Skip to main content

Full text of "The existence of subgame-perfect equilibrium in games with simultaneous moves"

See other formats


Digitized  by  the  Internet  Archive 

in  2011  with  funding  from 

Boston  Library  Consortium  Member  Libraries 


http://www.archive.org/details/existenceofsubgaOOharr 


HB31 
.M415 
no.  S'U 


f  economics 


1991 


THE  EXISTENCE  OF  SUBGAME- PERFECT  EQUILIBRIUM 
IN  GAMES  WITH  SIMULTANEOUS  MOVES 


Christopher  Harris 


Number  570 


December  1990 


massachusetts 
institute  of 

technology 

- 


50  memorial  drive 
Cambridge,  mass. 02139 


THE  EXISTENCE  OF  SUBGAME- PERFECT  EQUILIBRIUM 
IN  GAMES  WITH  SIMULTANEOUS  MOVES 


Christopher  Harris 


Number  570  December  1990 


ft-'.. 


M.l.T.  LiBi 

apr  o  1  nm 


'  '■  ■  fl "-'      Iffl 


The  Existence  of  Subgame— Perfect  Equilibrium 

in  Games  with  Simultaneous  Moves: 

a  Case  for  Extensive— Form  Correlation 


Christopher  Harris 
Nuffield  College,  Oxford 

December  1990^ 


Abstract 

The  starting  point  of  this  paper  is  a  simple,  regular,  dynamic  game  in  which 
subgame-perfect  equilibrium  fails  to  exist.  Examination  of  this  example  shows  that 
existence  would  be  restored  if  players  were  allowed  to  observe  public  signals.  The  main 
result  of  the  paper  shows  that  allowing  the  observation  of  public  signals  yields  existence, 
not  only  in  the  example,  but  also  in  an  extremely  large  class  of  dynamic  games. 


I  should  like  to  thank  Drew  Fudenberg,  Daniel  Maldoom,  Andreu  Mas— Colell,  Eric 
Maskin,  Jean— Francois  Mertens,  Meg  Meyer,  Jean  Tirole,  John  Vickers,  and 
participants  at  seminars  and  workshops  at  Oxford,  Stonybrook  and  Harvard  for 
helpful  comments  on  the  earlier  paper  from  which  the  present  paper  is  drawn. 
The  present  paper  is  drawn  from  an  earlier  paper  entitled  "The  Existence  of 
Subgame— Perfect  Equilibrium  with  and  without  Markov  Strategies:  a  Case  for 
Extensive  Form  Correlation". 


1    Introduction 

Consider  the  following  example  of  a  dynamic  game.  Firms  set  out  with  exogenously 
specified  capacities  (which  are  known  to  all).  In  period  one  they  simultaneously  choose 
investment  levels  (possibly  on  a  random  basis),  and  are  then  informed  of  one  another's 
choices.   The  result  is  a  change  in  capacities.   In  period  two  firms  simultaneously  choose 
production  levels  (within  their  capacity  constraints),  and  are  then  informed  of  one 
another's  choices.   The  result  is  a  change  in  inventories.  In  period  three  firms 
simultaneously  choose  prices.  The  vector  of  prices  chosen  affects  the  vector  of  demands  for 
their  products,  but  so  do  certain  exogenous  random  factors.   Firms  are  informed  of  one 
another's  chosen  prices  and  of  the  final  realised  demands.   (The  demands  bring  about  a 
second  change  in  inventories.)  The  three  period  cycle  of  choices  the  begins  afresh.   And  so 
on. 

This  game  is  an  example  of  a  game  of  the  following  general  type.   Time  is  discrete. 
There  is  a  finite  number  of  active  players.   There  is  also  a  passive  player,  Nature.  In  any 
given  period:  all  players  (both  active  and  passive)  know  the  outcomes  of  all  previous 
periods;  the  set  of  actions  available  to  any  active  player  is  compact,  and  depends 
continuously  on  the  outcomes  of  the  previous  periods;  the  distribution  of  Nature's  action 
(which  is  given  exogenously)  depends  continuously  on  the  outcomes  of  the  previous  periods; 
the  players  (active  and  passive)  choose  their  actions  simultaneously;  and  the  outcome  of 
the  period  is  simply  the  vector  of  actions  chosen.   The  outcome  of  the  game  as  a  whole  is 
the  (possibly  infinite)  sequence  of  outcomes  of  all  periods,  and  players'  payoffs  are  bounded 
and  depend  continuously  on  the  outcome  of  the  game. 

This  is  as  regular  a  class  of  dynamic  games  as  one  could  ask  for.   A  counter-example 
shows,  however,  that  games  in  this  class  need  not  have  a  subgame— perfect  equilibrium.   It 
is  therefore  necessary  to  extend  the  equilibrium  concept  in  such  a  way  that  existence  is 
restored. l 


The  problem  is  reminiscent  of  that  of  the  non-existence  of  Nash  equilibrium  in  pure 


What  is  the  most  natural  extension  of  the  equilibrium  concept?   One  clue  as  to  the 
answer  to  this  question  is  provided  by  the  following  considerations.   First,  if  we  simplify 
the  class  of  games  under  consideration  by  requiring  that  players'  action  sets  are  always 
finite,  then  subgame— perfect  equilibrium  always  exists.   Hence  the  non-existence  problem 
appears  to  relate  specifically  to  the  fact  that  we  allow  a  continuum  of  actions.   Secondly, 
suppose  instead  that  we  simplify  the  class  of  games  under  consideration  by  assuming  that 
players'  action  sets  are  independent  of  the  outcomes  of  previous  periods.  Then  a  natural 
way  of  approximating  a  game  is  to  consider  subsets  of  players'  action  sets  that  consist  of 
large  but  finite  numbers  of  closely  spaced  actions.2  Moreover,  if  one  takes  a  sequence  of 
increasingly  fine  approximations,  and  a  subgame— perfect  equilibrium  of  each  of  the 
approximations,  then  it  is  natural  to  expect  that  any  limit  point  of  the  sequence  of 
equilibrium  paths  so  obtained  will  be  an  equilibrium  path  of  the  original  game.3  One  can, 
however,  find  examples  in  which  the  limit  point  involves  a  special  form  of  extensive— form 
correlation:  at  the  outset  of  each  period  agents  observe  a  public  signal.4 


strategies.   But  there  is  one  important  difference:  if  agents'  action  sets  are  always  finite, 
then  subgame— perfect  equilibrium  does  exist. 

2  Hell  wig  and  Leininger  (1986)  were  the  first  to  introduce  such  an  approximation,  in 
the  context  of  finite— horizon  games  of  perfect  information.  They  proved  an 

upper— semicontinuity  result:  they  showed  that  any  limit  point  of  equilibrium  paths  of  the 
finite  approximations  is  an  equilibrium  path  of  the  original  game.   Their  result  can, 
however,  be  understood  in  terms  of  the  results  of  Harris  (1985a).  Indeed,  in  that  paper  it  is 
shown  that  the  point— to— set  mapping  from  period— t  subgames  of  a  game  of  perfect 
information  to  period— t  equilibrium  paths  is  upper  semicontinuous.   Hence,  in  order  to 
obtain  their  result,  one  need  only  introduce  a  dummy  player  who  chooses  n  e  IN  U  {qd}  at 
the  outset  of  the  game.  If  n  <  od  then  the  remaining  players  will  be  restricted  to  actions 

chosen  from  the  n     approximation  to  their  action  sets.  If  n  =  od  then  they  will  be  free  to 
choose  actions  from  their  original  action  sets. 

3  Convergence  of  equilibrium  paths,  as  used  implicitly  in  Harris  (1985a)  and  explicitly 
in  Hellwig  and  Leininger  (1986)  and  Borgers  (1989)  seems  more  relevant  than  the 
convergence  of  strategies  considered  by  Fudenberg  and  Levine  (1983)  and  Harris  (1985b). 

4  Fudenberg  and  Tirole  (1985)  consider  two  such  games.  They  point  out  that  the 
obvious  discretisations  of  these  games  have  a  unique  symmetric  subgame— perfect 
equilibrium,  and  that  the  limiting  equilibria  obtained  as  the  discretisation  becomes 
arbitrarily  fine  involve  correlation.   Their  games  do  not,  however,  yield  counterexamples  to 
the  existence  of  subgame— perfect  equilibrium.   For  both  games  possess  asymmetric 
equilibria  which  do  not  involve  correlation. 


These  considerations  suggest  that,  at  the  very  least,  the  equilibrium  concept  should 
be  extended  to  allow  the  observation  of  public  signals.  This  minimal  extension  is  also 
sufficient.   For,  first,  it  restores  existence.   Secondly,  with  it,  the  equilibrium 
correspondence  of  a  continuous  family  of  games  is  upper  semicontinuous.   In  particular,  any 
limit,  point  of  equilibrium  paths  of  finite  approximations  to  a  game  whose  action  sets  are 
continua  is  an  equilibrium  path  of  the  game.   Thirdly,  any  limit  point  of  subgame-perfect  e- 
equilibrium  paths  of  a  game  is  an  equilibrium  path  in  the  extended  sense. 5 

The  basic  structure  of  the  proof  of  the  existence  of  subgame-perfect  equilibrium  in 
the  extended  sense  is  the  same  as  the  structure  of  the  proof  of  the  existence  of 
subgame-perfect  pure— strategy  equilibrium  in  games  of  perfect  information  given  in  Harris 
(1985a).  It  breaks  down  into  two  main  parts,  the  backwards  and  the  forwards  programs. 
The  backwards  program  is  most  easily  explained  in  the  context  of  a  game  with  finite 
horizon  T.  In  such  a  game,  it  consists  in  solving  recursively  for  what  turn  out  to  be  the 
equilibrium  paths  of  the  game.   More  precisely,  one  solves  first  for  the  equilibrium  paths  of 
period— T  subgames.   (The  set  of  equilibrium  paths  of  a  period— T  subgame  is  the  convex 
hull  of  the  set  of  probability  distributions  over  period— T  outcomes.)  Then  one  solves  for 
the  equilibrium  paths  of  period— (T— 1)  subgames.   (The  set  of  equilibrium  paths  of  a 
period— (T— 1)  subgame  is  a  set  of  probability  distributions  over  period— (T— 1)  histories,  i.e. 
a  joint  distribution  over  period— (T— 1)  and  period— T  outcomes.)   And  so  on  until  period  1 
is  reached,  and  a  set  of  probability  distributions  over  period— 1  histories  is  obtained. 

There  is  of  course  no  general  guarantee  that  the  paths  obtained  in  the  course  of 
carrying  out  the  backwards  program  really  are  equilibrium  paths  until  equilibrium 
strategies  generating  them  have  been  constructed.  To  construct  such  strategies  is  the 
purpose  of  the  forwards  program.   The  essential  idea  is  this.   Pick  any  of  the  period— 1 
paths  obtained  as  the  final  product  of  the  backwards  program.   Such  a  path  is  a  probability 
distribution  over  period— 1  histories.  Its  marginal  over  period— 1  outcomes  can  be  used  to 


5  Chakrabarti  (1988)  claims  that  subgame-perfect  e-equilibria  exist.  We  believe  this 

claim,  but  are  unable  to  vouch  for  the  proof. 


construct  period— 1  behaviour  strategies  for  the  players.  Its  conditional  over  period— 2 
histories  yields  continuation  paths  to  be  followed  from  period  2  onwards  in  the  event  that 
players  do  not  deviate  from  their  period— 1  strategies.  In  order  to  obtain  continuation 
paths  in  the  event  that  some  player  does  deviate  from  his  period— 1  strategy,  it  is  sufficient 
to  choose  from  among  the  period— 2  paths  obtained  in  the  course  of  the  backwards  program 
some  path  which  minimises  the  continuation  payoff  of  that  player.  In  this  way 
continuation  paths  for  all  period— 2  subgames  are  obtained.   Period— 2  strategies  are  then 
obtained  from  the  marginals  of  these  paths  over  period— 2  outcomes.   And  so  on  until 
finally  period— T  strategies  are  obtained. 

Implementing  this  proof  does,  however,  involve  two  significant  new  complications. 
First,  for  technical  reasons,  it  is  necessary  to  maintain  the  induction  that  the  point— to— set 
mapping  from  period— t  subgames  to  period— t  equilibrium  paths  is  upper  semicontinuous. 
This  follows  from  a  generalisation  of  the  work  of  Simon  and  Zame  (1990).   Secondly, 
because  players  may  randomise,  it  is  essential  to  ensure  that  all  the  strategies  constructed 
are  measurable. 6  That  this  requirement  can  be  met  emerges  as  a  natural  corollary  of  the 
construction  employed:  one  can  always  construct  a  measurable  family  of  conditional 
distributions  from  a  measurable  family  of  probability  distributions.7 

The  organisation  of  the  paper  is  as  follows.   Section  2  sets  out  the  counterexample 
to  the  existence  of  subgame— perfect  equilibrium.   This  example  involves  a  three— period 
game  with  two  players  in  each  period.  In  this  game,  each  player's  action  set  is  compact 


6  Hellwig  and  Leininger  (1987)  were  the  first  to  draw  attention  to  the  desirability  of 
ensuring  that  strategies  are  measurable.  Their  approach  to  the  existence  of 
subgame— perfect  equilibrium  is  not,  however,  necessary  in  order  to  obtain  measurability. 
Indeed,  if  the  assumptions  of  Harris  (1985a)  are  specialised  appropriately  (by  assuming 
that  the  embedding  spaces  for  players'  action  sets  are  compact  metric  instead  of  compact 
Hausdorff),  then  it  is  simple  to  show  that  exactly  the  construction  given  there  can  be 
carried  out  in  a  measurable  way.  Indeed,  in  the  notation  of  Harris  (1985a;  p. 625),  all  that 

is  necessary  is  to  ensure  that  the  functions  gj   can  be  chosen  to  be  measurable;  and  that 

this  is  possible  follows  from  the  argument  given  in  the  first  paragraph  of  the  proof  of 
Lemma  4.1  in  the  present  paper. 

7  Had  we  employed  a  dynamic— programming  approach,  the  problem  of  ensuring 
measurability  of  strategies  would  have  been  much  harder  —  see  below  for  further  discussion. 
Indeed,  we  do  not  know  how  to  solve  the  measurable  selection  problem  that  arises  when 
such  an  approach  is  employed. 


and  independent  of  the  outcomes  of  previous  periods,  and  each  player's  payoff  function  is 
continuous.   Section  3  formulates  the  basic  framework  used  in  this  paper  for  the  analysis  of 
dynamic  games.   Section  4  shows  that  when  the  dynamics  are  continuous,  players'  payoff 
functions  are  continuous,  and  the  framework  is  extended  to  allow  the  observation  of  public 
signals,  subgame— perfect  equilibrium  exists.   This  is  the  main  result  of  the  paper.   Section 
4.1  provides  a  detailed  overview  of  the  proof.   Section  4.2  develops  the  analysis  necessary 
for  the  backwards  program.   Section  4.3  develops  the  analysis  necessary  for  the  forwards 
program.   Section  4.4  exploits  the  analysis  of  Sections  4.2  and  4.3  to  prove  the  main  result, 
showing  in  particular  how  to  deal  with  the  infinite— horizon  case.   Section  5  attempts  to 
develop  a  perspective  on  the  main  result.  It  is  shown  there  that  there  are  at  least  three 
types  of  game  in  which  the  introduction  of  public  signals  is  not  necessary  to  obtain  the 
existence  of  subgame-perfect  equilibrium:  games  of  perfect  information,  finite-action  games, 
and  zero-sum  games.   It  is  also  shown  that  the  introduction  of  public  signals  accommodates 
all  the  equilibria  that  one  can  obtain  by  taking  finite-action  approximations  to  a  game,  or 
by  considering  e— equilibria  of  a  game.  We  do  not  know  whether  the  introduction  of  public 
signals  is  the  minimal  extension  with  these  two  properties,  but  it  seems  plausible  to  argue 
that  it  is. 


2    The  Counterexample 

A  counterexample  to  the  existence  of  subgame— perfect  equilibrium  in  the  class  of 
dynamic  games  described  in  the  introduction  must  have  a  certain  minimum  complexity. 
First  of  all,  it  must  have  at  least  two  periods.   Otherwise  the  standard  existence  theorem 
for  Nash  equilibrium  would  apply.   Secondly,  there  must  be  at  least  two  players  in  the 
second  period.   For  with  only  one  player  in  the  second  period,  the  correspondence 
describing  the  continuation  payoffs  available  to  the  player  or  players  in  period  one  would 
be  convex  valued.   (The  player  in  period  two  can  randomise  over  any  set  of  alternatives 
between  which  he  is  indifferent.)   So  the  existence  result  of  Simon  and  Zame  (1990)  would 
apply  to  period  one.   Thirdly,  there  must  be  at  least  two  players  in  period  one.  For  the 
continuation— payoff  correspondence  for  period  one  will  be  upper  semicontinuous,  and  a 
single  player  will  therefore  be  able  to  find  an  optimum. 

The  counterexample  presented  in  this  section  is  not  this  minimal.  It  has  three 
periods,  with  two  players  in  each  period.   It  would  be  of  considerable  interest  to  find  an 
example  with  only  two  periods  and  two  players  in  each  period.   For,  aside  from  being 
minimal  in  relation  to  the  class  of  dynamic  games  considered  in  this  paper,  such  an 
example  would  settle  the  question  as  to  whether  equilibrium  exists  in  two— stage  games  or 
not. 

The  cast  of  the  counter— example,  and  their  choice  variables,  are  as  follows.  In 
period  one  two  punters  A  and  B  pick  a  6  [0,1]   and  b  €  [0,1]   respectively.   In  period 
two,  two  greyhounds  C  and  D  each  receive  an  injection  of  size  a  +  b  ,  which  changes 
their  attitude  to  the  race.   They  pick  c  e  [0,1]   and  d  6  [0,1]  ,  which  are  the  times  in  which 
they  complete  the  course.  In  period  three  each  of  two  referees  E  and  F  must  declare  a 
winner,  picking  e  6  {C,D}  and  f  e  {C,D}  respectively.  In  each  period  choices  are  made 
simultaneously,  and  players  in  later  periods  observe  the  actions  taken  by  players  in  earlier 
periods. 


Punter  A  obtains  a  payoff  of  1  —  a  if  greyhound  C  is  declared  the  winner  by 
both  referees,  and  — 1  —  a  otherwise.   Similarly,  punter  B  obtains   1  —  b  if  both  referees 
declare  D  to  be  the  winner,  and  — 1  —  b  otherwise.  In  other  words,   A  wants  C  to  win, 
B  wants  D  to  win,  and  both  want  a  result.   They  would  also  like  to  keep  their 
contributions  to  the  injection  as  small  as  possible.   The  payoff  to  greyhound  C  is  2c  if 
e  =  C  and  1  —  (a  +  b)(l  —  c)  if  e  =  D  .   That  is,  the  form  of  his  payoff  depends  on 
whether  he  or  the  other  greyhound  is  declared  the  winner  by  referee  E  ,  but  either  way  he 
would  prefer  to  run  the  race  as  slowly  as  possible.   Also,  he  would  prefer  to  be  first  rather 
than  second  provided  that  he  does  not  have  to  run  too  fast.   The  payoff  to  greyhound  D  is 
2d  if  e  =  D  and  1  —  (a  +  b)(l  —  d)  if  e  =  C  :  like  greyhound  C  ,  he  is  only  interested 
in  the  verdict  of  referee  E  .   Lastly,  referee  E  gets  payoff  d  is  he  declares   C  to  be  the 
winner  and  c  if  he  declares  D  to  be  the  winner.   Referee  F  's  payoffs  are  identical. 

It  will  be  seen  that,  in  this  game,  each  player's  action  set  is  compact  and 
independent  of  the  actions  chosen  by  earlier  players,  and  that  players'  payoffs  are 
continuous. 

The  game  centres  around  the  two  greyhounds.  Because  referee  E  chooses  C  when 
c  <  d  and  D  when  c  >  d  ,  the  race  between  them  is  essentially  a  game  of  timing  with 
the  discontinuous  payoffs  (2c,  1  —  (a  +  b)(l  —  d))  if  c  <  d  ,   (1  -  (a  +  b)(l  -  c),  2d)  if 
c  >  d  ,  and  some  convex  combination  of  these  two  payoffs  when  c  =  d  .   (Note  that  the 
weights  in  the  convex  combination  can  depend  on  (c,d)  .)   Standard  considerations 
therefore  yield  the  following  lemma. 

Lemma  2.1    Suppose  that  a  +  b  >  0  .  Then  both  C  and  D  use  mixed  actions  with 
distribution  function  G  given  by  G(x)  =  0  for  x  e  [0,|]   and  G(x)  =  (2x  -  l)/(2x  -  1 
+  (a  +  b)(l-x))  for  xe  [$,1]  .       □ 

Notice  that  this  mixed  action  is  non-atomic,  that  its  support  is   [j,l]  ,  and  that  all 
its  probability  mass  concentrates  at  £  as  a  +  b  -»  0+  .   Not  surprisingly,  then,  we  obtain: 


Lemma  2.2    Suppose  that  a  +  b  =  0  .   Then  both  C  and  D  choose  $  with  probability 
one.        d 

The  remainder  of  the  argument  is  straightforward.  If  a  +  b  >  0  then  c  will  equal 
d  with  probability  zero.   Hence  both  referees  will  always  agree.   Also,  each  greyhound  will 
win  exactly  half  the  time.   The  payoffs  to  A  and  B  are  therefore  —a  and  — b.  If,  on  the 
other  hand,  a  +  b  =  0  ,  then  c  =  d  =  \  with  probability  one.   Hence  each  referee  is 
indifferent  as  to  which  greyhound  he  declares  to  be  the  winner.   Suppose  that  E  opts  for 
C  with  probability  p  and  D  with  probability  1  —  p  ,  and  that  F  opts  for  C  with 
probability  q  and  D  with  probability   1  -  q  .  Then  A  's  payoff  is  2pq  —  1  and  B  's  is 
2(l-p)(l-q)-l. 

It  is  easy  to  see,  however,  that  the  game  with  these  payoffs  between  A  and  B  has 
no  equilibrium.  Indeed,  any  a  >  0  is  strictly  dominated  for  A  ,  as  is  any  b  >  0  for  B  . 
So  the  only  possibility  for  an  equilibrium  is  for  A  to  set  a  =  0  with  probability  one  and 
B  to  set  b  =  0  with  probability  one.   But  if  b  =  0  then  it  must  be  that  2pq  —  1  >  0 
—  otherwise  A  would  raise  a  from  zero.   Similarly,  2(1  —  p)(l  —  q)  —  1  >  0  .   And  these 
two  inequalities  are  mutually  inconsistent.8  This  contradiction  establishes  the  counter- 
example. 

The  essence  of  the  counterexample  is  this.   As  long  as  a  +  b  >  0  ,  both  greyhounds 
use  strictly  mixed  actions.   This  behaviour  on  their  part  generates  a  public  signal 
endogenously  within  the  game.   The  two  referees  use  this  public  signal  to  co-ordinate  their 
actions.  When  a  +  b  =  0  ,  however,  this  signal  suddenly  degenerates,  and  the  only  way 
in  which  the  referees  can  coordinate  their  actions  is  by  both  choosing  C  with  probability 
one  or  both  choosing  D  with  probability  one.   But  if  they  both  choose  C  then  B  gets 


8  To  see  this,  note  that  the  sum  of  the  payoffs  of  the  two  players  A  and  B  is  0  if 

e  =  f  and  —1  if  e  ^  f.   Hence  the  expected  value  of  this  sum  is  non— positive.   If  the 
expected  value  is  strictly  negative,  then  the  payoff  of  at  least  one  player  is  also  strictly 
negative.  If  the  expected  value  of  the  sum  is  zero,  then  e  and  f  must  be  perfectly 
correlated.  In  this  case  the  payoff  of  the  player  against  whom  the  decision  goes  is  — 1. 


—1  .   So  he  would  prefer  the  truly  random  outcome.   Similarly,  if  both  referees  choose  D 
then  A  will  wish  to  restore  the  truly  random  outcome. 

In  the  light  of  this  explanation,  it  is  natural  to  try  to  restore  existence  by  allowing 
players  to  observe  suitable  public  signals.  In  the  present  example,  this  would  amount  to 
allowing  the  two  referees  to  toss  a  coin  to  determine  the  winner  in  the  event  of  a  tie  (which 
would  certainly  restore  existence).  That  such  an  extension  yields  existence  in  the  general 
case  will  be  demonstrated  in  Section  4. 


10 


3    The  Basic  Framework 

3.1        The  Data 

There  is  a  non-empty  finite  set    J<A  of  active  players,  indexed  by  i  or  j  .  There 
is  also  a  passive  player,  Nature,  whose  index  is  i  or  j  =  0  .  The  overall  set  of  players  is 
therefore    J=  J^i  U  {0}  .   Time  is  discrete,  and  is  indexed  by  t  or  se  J"=  {0,1,2,...}  . 
For  notational  reasons,  we  assume  that    J  n  «J=  0  .   The  players  play  one  of  a  family  of 
games,  parameterised  by  x    eX    .   In  each  active  period  t  6  STjd  =  {1,2,...}  each  player 
i  e  J  must  choose  an  action.  The  set  of  actions  available  to  her  or  him  depends  on  which 
game  is  being  played,  and  on  the  outcomes  of  previous  periods.   This  situation  is  modelled 
by  a  point— to— set  mapping  A..:  X  —   -»  Y,.  .   The  vector  x  —    e  X  —    lists  the 
parameter  of  the  game  being  played,  and  the  outcomes  of  any  preceding  periods,  while 
A,  .(x      )  C  Y,  -  is  the  set  of  actions  available.   For  reasons  of  expositional  economy,  we 
take  it  that   A.^x  —  )  =  Y,^  for  all  x  —    €  X  ~~    .9  The  set  of  outcomes  possible  in 
period  t  is  nothing  more  than  the  set  of  profiles  of  actions  that  players  can  take.   It  is 
modelled  by  the  point-to-set  mapping  A .:  X  —   -♦  Y.  ,  where  Y.  =  x       Y..   and 
A,(x      )  =  x       A,.(x      )  for  all  x       EX        .   And  X    is  simply  the  set  of  all  pairs 
(x  ~  ,y, )  such  that  x  —   eX~    and  y,  e  A,(x  —  )  ,  i.e.  the  graph  of  A,  .  Finally,  the 
payoff  of  any  active  player  i  depends  on  which  game  x    e  X     is  played,  and  the 
outcomes  of  all  periods.   It  is  modelled  by  a  function  u-:  Xw  -» IR  .   Here  X00  is  the  set  of 


,0 


t-1  /  N    .    Vt-1 


vectors  x  =  (x0,Xp...)  e  X    x    x       Y.     such  that  x  —    =  (xn,x.  ,..,x,_.)  eX~    for  all 
t  >  1  .   (Nature  does  not  have  a  payoff  function.) 

We  make  the  following  standing  assumptions  about  these  data: 
(i)  X    ,  and  all  of  the  sets  Y..  ,  are  non-empty  complete  separable  metric  spaces; 

(ii)         for  all  i  €  X4  and  all  t  >  1  ,  A..:  X  —   -♦  J6(Y,-)  is  measurable; 
(iii)        for  all  i  e  J<A  ,  u-  is  bounded  and  measurable. 


9  This  assumption  does  in  principle  involve  a  slight  loss  of  generality.   However,  since 

we  shall  in  any  case  fix  Nature's  strategy  below,  there  seems  to  be  little  advantage  in 
constraining  her  actions  as  well. 


11 


(In  this  paper,  'measurable'  will  always  mean  Borel  measurable  unless  explicitly  stated  to 
the  contrary;  and   •%(•)  will  always  denote  the  space  of  non-empty  compact  subsets  of  • 
endowed  with  the  Hausdorff  metric.)  These  assumptions  are  designed  to  ensure  that  every 
player  possesses  at  least  one  strategy,  and  that  associated  with  every  strategy  profile  and 
every  subgame  there  is  a  well  defined  payoff  for  every  active  player  (see  below).   Stronger 
assumptions  are  needed  to  ensure  the  existence  of  equilibrium. 

3.2        Strategies  and  extended  strategies 

In  the  standard  version  of  our  model,  the  evolution  of  the  game  is  as  follows.   At  the 
beginning  of  period  1,  players  are  told  which  game  they  are  playing.   That  is,  they  are 
informed  of  x    .   They  then  simultaneously  choose  actions  from  their  action  sets  for  period 
1.   This  completes  period  1.   At  the  beginning  of  period  2,  players  are  told  the  outcome  of 
period  1.   That  is,  they  are  told  y,  .   They  then  choose  simultaneously  from  their  action 
sets  for  period  2.   This  completes  period  2.   And  so  it  goes  on.  It  follows  that  players' 
information  in  period  t  can  be  summarised  by  a  vector  x  —    eX~    ,  and  that  X  —     can 
be  identified  with  the  set  of  subgames  in  period  t  . 

Definition  3.1    For  each  t  >  1  and  each  i  e  J  ,  a  strategy  for  player  i  in  period  t  is  a 

measurable  function  f,.:  X  ~~   ->  9Jl^{4-)  such  that  supp[f,.(x  _  )]  C  A,.(x      )  for  all 

t-1  ,  vt-l 
x        6  X 

Here   9Ji{XA  denotes  the  set  of  probability  measures  over  Y,.  .   (In  this  paper, 
'probability  measure'  will  always  mean  Borel  probability  measure;  and  spaces  of 
probability  measures  will  be  taken  to  be  endowed  with  the  weak  topology10  unless  explicitly 
stated  to  the  contrary.) 


10         In  the  terminology  of  Parthasarathy  (1967),  a  sequence  of  probability  measures 

{A  }  C  &M(Y.)  converges  in  the  weak  topology  to  A  iff  Jx  dXn   converges  to  /xdA  for 

all  continuous  bounded  x-  Yf  -» IR  . 


12 


The  f.-   are  behaviour  strategies.   For  each  t  ,  i  and  x        ,  f,  .(x      )   can  be 
thought  of  as  the  randomising  device  that  player  i  will  use  to  choose  among  the  actions 
available  to  him  in  period  t  when  x        is  the  previous  history  of  the  game.   As  such,  it  is 
independent  of  the  devices  used  by  the  other  players  in  period  t  ,  and  of  all  devices  used  in 
all  preceding  and  subsequent  periods. 

Strategies  and  strategy  profiles  certainly  exist  under  our  standing  assumptions. 
Moreover,  given  any  strategy  profile  and  any  subgame,  the  payoffs  of  active  players  can  be 
calculated  in  the  natural  way.  We  may  therefore  define  subgame— perfect  equilibrium  as 
follows. 


Definition  3.2  A  subgame— perfect  equilibrium  is  a  strategy  profile  <f,- 1 1  >  1,  i  e  Jj&> 
such  that,  for  all  t  >  1,  all  x  —  e  X  —  ,  and  all  i  G  X^"  ,  player  i  cannot  improve  his 
payoff  in  subgame  x        by  a  unilateral  change  in  his  strategy. 


As  the  counterexample  of  Section  2  shows,  however,  subgame— perfect  equilibrium 
need  not  exist  in  general.   We  therefore  introduce  the  concept  of  an  extended  strategy. 
Suppose  that   {£.  |t  >  1}  is  a  sequence  of  signals,  drawn  independently  from   [0,1] 
according  to  the  uniform  distribution.   Suppose  that,  at  the  beginning  of  each  active  period 
t  >  1  ,  players  are  informed  not  only  of  the  outcome  of  the  preceding  period,  but  also  of 
£,  .  Then  the  information  available  to  them  when  they  make  their  choices  can  be 
summarised  by  a  vector  h  —   =  (x  —  ,£  )  ,  where   £   =  (£-,,..., £t)  ■   We  refer  to  the  set 
H  ~~   =  X  —   x  [0,1]    of  such  vectors  as  the  set  of  extended  subgames  of  the  game  in 
period  t  . 

Definition  3.3  For  each  t  >  1  and  each  i  e  J  ,  an  extended  strategy  for  player  i  in 
period  t  is  a  measurable  function  f,.:  H  —  ->  &J((Y..)  such  that  supp[f,.(x  —  ,£  )] 
C  Ati(xt_1)  for  all  (xt_1,^)  e  Ht_1. 


13 

Once  again  the  f,.   are  behaviour  strategies.  So  f..(x  —  ,£  )  can  be  thought  of  as 
the  randomising  device  that  player  i  will  use  to  choose  among  the  actions  available  to  him 
in  period  t  ,  when  x  —    is  the  previous  history  of  the  game  and   £    is  the  vector  of  public 
signals  observed  up  to  and  including  period  t  .  But  this  time  this  device  is  independent, 
not  only  of  the  devices  used  by  the  other  players  in  period  t  and  of  all  devices  used  in  all 
other  periods,  but  also  of  all  the  public  signals. 

Definition  3.4  A  subgame— perfect  equilibrium  in  extended  strategies  is  an  extended- 
strategy  profile  <f.  |t  >  1,  i  e  X^>   such  that,  for  all  t  >  1  ,  all  h  —    eH~    ,  and  all  i 
6  Jj6  ,  player  i  cannot  improve  his  payoff  in  extended  subgame  h         by  a  unilateral 
change  in  his  strategy. 

A  subgame— perfect  equilibrium  in  extended  strategies  allows  a  limited  amount  of 
coordination:  after  the  outcome  of  a  period  has  been  realised,  the  players  can  coordinate  on 
the  continuation  equilibrium  to  be  played  following  that  outcome  by  exploiting  the  next 
public  signal.  It  is  therefore  a  limited  kind  of  extensive— form  correlated  equilibrium.   (The 
correlation  is  limited  in  that  the  players  observe  a  common  signal,  rather  than  each 
privately  observing  a  single  component  of  a  vector  of  signals.) 


14 


4    The  Main  Result 


We  need  the  following  assumptions: 
(Al)     for  all  t  >  1  and  all  i  e  J^i  ,  the  mapping  A.-:  X  ~~   -»  J£(T..)  is  continuous; 
(A2)     for  all  t  >  1  ,  a  strategy  f.*   (and  not  an  extended  strategy)  for  Nature  in  period  t 

is  given,  and  f,*:  X  —   -*  9M (Y,,-.)  is  continuous; 
(A3)     for  all  i  G  J^i  ,   u-   is  bounded  and  continuous. 

Taken  in  conjunction  with  the  standing  assumptions  made  in  respect  of  the  basic 
framework,  they  guarantee  the  existence  of  a  subgame— perfect  equilibrium  in  extended 
strategies. 

The  present  section  is  devoted  to  a  proof  of  this  fact.  It  begins,  in  Section  4.1,  with 
an  overview  of  the  proof.  The  proof  involves  two  main  steps.   The  analysis  necessary  for 
the  first  of  these,  namely  the  backwards  program,  is  given  in  Section  4.2.   The  analysis 
necessary  for  the  other,  namely  the  forwards  program,  is  given  in  Section  4.3.   Section  4.4 
then  integrates  the  backwards  and  the  forwards  programs  to  complete  the  proof.   The 
reader  may  wish  to  read  Section  4.1  and  then  skip  to  Section  5. 

4.1       Overview 

Imagine  for  the  purposes  of  the  present  subsection  that  the  game  has  horizon  T  . 
Then  the  proof  divides  into  essentially  two  steps.  In  the  first  step,  one  programs  backwards 
from  stage  T  ,  finding  what  turn  out  to  be  the  equilibrium  paths  of  the  game.  In  the 
second,  one  programs  forwards,  constructing  strategies  that  generate  these  paths  and  that 
constitute  an  equilibrium. 

The  basic  ideas  of  the  backwards  program  are  as  follows.   First,  for  each  history 

T— 1        T— 1 
x         e  X         ,  find  the  set  of  Nash  equilibria  (including  Nash  equilibria  in  mixed 

strategies)  that  can  occur  in  the  final  stage  of  the  game.   Let   C™(x        )  be  the  set  of 

probability  distributions  over  Y™  that  result  from  these  Nash  equilibria.   Because  action 

T— 1 
sets  are  compact  and  payoff  functions  are  continuous,   C™(x       )  is  a  non-empty  compact 


15 


T— 1 
set  contained  in    9JC{Xrr)  .   Because  actions  sets  depend  continuously  on  x         ,  the 

T— 1 
point— to— set  mapping  C™  from  X  to     ^(^^(Yrp))  is  upper  semicontinuous.11 

•p o 

Next,  let  x  be  the  history  of  the  game  prior  to  period  T— 1  .   For  any  v™  - 

T— 2 
G  AT_1  (x       )  ,  the  set  of  continuation  paths  consistent  with  equilibrium  is  given  by 

<p o 

co[CT(x        ,yrp_1 )]  .   (This  set  consists,  in  effect,  of  those  probability  distributions  over 

YT  that  can  be  obtained  by  randomising  over  Nash  equilibria  in  Crp(xrj,_, )  .)   For  each 

selection  cT(x        ,  • )  from  the  correspondence  co[Cy(x        ,  • )]  ,  find  the  set  of  Nash 

equilibria  that  can  occur  in  period  T— 1  when  the  continuation  of  the  game  is  specified  by 

^p 2  T 2 

Crp(x        , •)  .   (For  a  general  c,y(x        ,•)  ,  this  set  may  be  empty.   But  there  will  always 

<p 2 

be  at  least  one  c,y,(x        ,•)  for  which  it  is  not.   This  is  essentially  the  content  of  the 

existence  theorem  of  Simon  and  Zame  (1990).)  And  for  each  Nash  equilibrium  associated 

T— 2 

with  Crp(x        ,  • )  ,  find  the  probability  distribution  over  Y™,  *  Y™  that  results  when 

that  equilibrium  is  played  in  period  T— 1   and  the  continuation  paths  are  given  by 

p<_2  "p 2 

Crp(x        ,  • )  .   Let   Crp_-.  (x        )  be  the  set  of  such  distributions  obtained  when  both 

T— 2 
Crp(x        , •)  ,  and  the  Nash  equilibrium  associated  with  it,  are  allowed  to  vary.   Then 

T— 2 
Cy,_|(x       )  is  a  non-empty  compact  subset  of   &J£(Yrr< _i*Yrp)  ,  and  the  point— to— set 

p 2 

mapping  C™,    from  X         to    ^(^^(Y^ixYrp))  is  upper  semicontinuous.   (This  is 

essentially  the  content  of  the  generalisation  of  Simon  and  Zame's  theorem  given  in 

Section  4.2.) 


11  The  relevance  of  the  assumption  that  players'  action  sets  in  period  T  depend 

continuously  on  the  history  of  the  game  prior  to  period  T  emerges  clearly  in  the  context 

of  a  two— period  game.   For  each  n  >  1  ,  let  a?  be  a  Nash  equilibrium  in  period  2  of  such  a 

game  when  the  outcome  of  period  1  is  y,  ,  and  suppose  that   (y^a^) -» (y^a^)   as  n  -»  <d  . 

Because  player  i  's  action  set  is  upper  semicontinuous  in  the  outcome  of  period  1,  the 
action  a„j  is  actually  available  to  him  in  period  2  when  the  outcome  of  period  1  is  y-,  . 

Also,  because  his  action  set  in  period  2  is  lower  semicontinuous  in  the  outcome  of  period  1, 
any  deviation  open  to  him  in  period  2  when  the  outcome  of  period  1  is  y,    can  be 

approximated  by  deviations  open  to  him  when  the  outcome  of  period  1  is  y,  .   Hence  a^. 
really  is  an  optimal  action. 


16 

Finally,  iterate  until  period  1  is  reached,  and  a  set  C.(x  )  of  probability 

T  0        0  0 

distributions  over   X...Y.   is  obtained  for  all  x    eX    .   (The  set  co[C,(x  )]  will  turn 

out  to  be  the  set  of  equilibrium  paths  of  the  game  x    .)  This  completes  the  backwards 

program. 

Consider  now  the  forwards  program.   Suppose  that,  for  each  x    6  X    ,  we  are  given 

c,(x  )  6  co[C.(x  )]  .   Since  any  path  in  co[C,(x  )]  can  be  viewed  as  a  randomisation  over 

paths  in  C.(x  ),  we  can  associate  a  probability  measure  d,(x  )  over  C,(x  )   with 

c..(x  )  .   Also,  for  each  d,(x  )  we  can  find  a  random  variable  e,(x  ,•):  [0,1]  -»  C,(x  ) 

whose  distribution  over  C,(x  )  is  d,(x  )  when  [0,1]  is  given  Lebesgue  measure.   For 

each  i  and  each   £    e  [0,1]  ,  let  f,-(x  ,£  )  be  the  marginal  of  e,(x  ,£  )  over  Y-.  .   By 

definition  of  C,(x  ),  there  exists  a  selection  c~(x  ,-,£  )  from  co[C2(x  ,-)]   such  that: 

(i)  the  f,-(x  ,£  )   constitute  a  Nash  equilibrium  in  period  1  when  the  continuation  paths 

are  given  by  c«(x  ,-,£  )  ;  and  (ii)  for  all  y,,  c«(x  ,y-.,£  )  is  the  conditional  distribution 

of  e,  (x  ,£  )  given  that  y1   is  the  outcome  in  period  1. 

Similarly,  beginning  with  cJx  ,£  )  6  co[C2(x  )]  one  obtains  d„(x  ,^  ),  eJx  ,(  ), 

12  2    2 

f„.(x  ,£  )  and  cJx  ,£  )  .  And  so  on  until  one  has  a  complete  set  of  strategies  f,-  for  all 

1  <  t  <  T  and  i  e  J^£  .  That  these  strategies  really  do  produce  the  paths  they  are 

intended  to  produce  is  primarily  a  technical  matter.   The  essential  points  are  that  the 

expectation  of  e,(x  -  ,£  -  ,-)  is  always  c,(x  —  ,f  —  )  ,  and  that  c,  ,  .(x  —  ,y,,^  )  is 

always  the  conditional  distribution  of  e,(x      ,£  )  given  y,  .   That  they  constitute  an 

equilibrium  follows  from  the  fact  that  the  f,-(x  —  ,£  )  constitute  a  Nash  equilibrium. 

This  ensures  that  no  single— period  deviation  is  profitable.  That  no  more  complex  deviation 

is  profitable  follows  from  a  standard  unravelling  argument.   This  completes  the  forwards 

program. 

Overall,  the  essential  ingredients  of  the  theorem  are  the  following:  players  action 

sets  are  compact;  players'  payoffs  are  continuous;  Nature's  strategy  is  continuous;  players' 

action  sets  in  a  given  period  depend  continuously  on  the  history  of  the  game  prior  to  that 

period;  players  observe  public  signals;  and  players'  payoffs  are  bounded.   The  role  played 


17 


by  the  first  three  of  these  should  be  obvious.  The  role  played  by  the  fourth  is,  roughly,  to 
ensure  that  the  point— to— set  mapping  from  subgames  in  period  t  to  equilibrium  paths  of 
those  subgames  is  upper  semicontinuous.  The  role  of  the  fifth  is  to  ensure  that  the 
point— to— set  mapping  from  subgames  in  period  t  to  equilibrium  paths  of  those  subgames  is, 
in  addition,  convex  valued.   And  the  sixth  is  needed  only  for  the  relatively  technical  reason 
that  we  do  not  insist  that  Nature's  behaviour  strategies  have  compact  support. 

4.2       The  backwards  program 

In  this  subsection  we  develop  the  analysis  necessary  for  the  backwards  program.   To 
this  end: 

(i)  let  X,  Y-  for  all  i  e  J ,  and  Z  be  non-empty  complete  separable  metric  spaces; 

(ii)         for  all  i  6  X4  ,  let  A-:  X  ->  Jfc(Y-)  be  a  continuous  point— to— set  mapping; 
(iii)         let  f*:  X  -»  9>Jt (Y*)   be  a  continuous  mapping; 
(iv)        let   Y  =  "jgjYj  and  let  A(x)  =  Y^  x    x    ^Aj(x)     for  all  x  £  X  ; 
(v)         let   C  :  gr(A)  -»  Jf(Z)  be  an  upper  semicontinuous  point— to— set  mapping; 
(vi)        for  each  i  e  J^i  ,  let  u-:  gr(C)  -»  R  be  a  continuous  function. 
The  similarity  of  this  notation  to  that  of  Section  3  is  deliberate,  and  is  intended  to  be 
helpful,  not  confusing. 

With  reference  to  the  model  of  Section  3  and  the  discussion  of  Section  4:  X  is  the  set 
of  past  histories;   A.   describes  the  dependence  of  player  i  's  action  set  on  the  past  history; 
f*   describes  Nature's  behaviour;   C  describes  the  set  of  possible  continuation  paths;  and 
u-  is  the  payoff  of  active  player  i  .  In  terms  of  a  somewhat  more  abstract  perspective, 
what  we  have  is  a  continuous  family  of  games  parameterised  by  x  e  X  .   The  three 
essential  ingredients  of  such  a  family  are:  the  continuity  of  the  action  sets;  the  upper 
semicontinuity  of  the  continuation  paths;  and  the  continuity  of  the  payoffs  in  the 
parameter,  the  outcome  and  the  continuation. 

For  each  x  6  X  we  are  interested  in  the  set  of  equilibrium  paths 
^C(x)  c  &Jt(Y  *  Z)  that  are  obtained  when  one  is  free  to  choose  any  randomisation  over 


18 


C(x,y)  as  the  continuation  path  following  the  outcome  y  G  A(x)  .   More  formally,   A 

€  Jc(x)  iff: 

(i)  the  marginal  \i  of  A  over  Y  is  a  product  measure  with  supp[/x]  C  A(x)  ; 

(ii)         the  marginal  /jg  of  \i  over  Y*  is  frt(*|x); 

(iii)        A  possesses  an  r.c.p.d.  (regular  conditional  probability  distribution)   v  over  Z 

such  that  supp[i/(  •  |  y)]  C  C(x,y)  for  all  y  e  A(x)  ; 
(iv)         moreover  the  marginals  \i-   of  (j,  over  the  Y-   constitute  a  Nash  equilibrium  when 

the  continuation  path  following  outcome  y  is  v(  •  |  y)  for  all  y  . 
(It  may  be  helpful  to  state  more  explicitly  what  is  meant  by  (iv).  Suppose  we  are  given 
probability  measures  fj,-   over  A.(x)  for  all  \£  J  and  a  transition  probability  v  such 
that  supp[^(-  |y)]  C  C(x,y)  for  all  y  6  A(x)  .   Let   A  be  the  measure  over  Y«Z 
obtained  by  combining  the  product  of  the  n-   with  v  .  Then  the  payoff  of  an  active  player 
i  is  /u-(x,y,z)dA(y,z)  ,  and  the  \i-   constitute  a  Nash  equilibrium  if  no  such  player  can 
improve  his  payoff  by  changing  his  probability  measure.) 

Lemma  4.1    #C  is  an  upper  semicontinuous  map  from  X  into  Jf( 3>JC{Y  *  Z))  . 

Proof    We  begin  with  some  notation.   For  each  i  ,  x  ,  and  y  6  A(x)  ,  let  p-(x,y) 
=  min{u.(x,y,z)  |  z  6  C(x,y)}  .  Then  p-(x,y)  is  the  lowest  payoff  that  can  be  imposed  on 
player  i  in  game  x  following  outcome  y  .  Standard  considerations  show  that  p.   is  lower 
semicontinuous.   Let   P:(x,y)  =  {z|z  e  C(x,y)   and  u-(x,y,z)  =  p-(x,y)}  .   We  shall  need  a 
measurable  selection  from  P.  .   Following  Deimling  (1985;  Proposition  24.3  and  Theorem 
24.3),  the  existence  of  such  a  selection  will  follow  if  P7  (D)  =  {(x,y)|D  n  P-(x,y)  j-  0}  is 
measurable  for  all  closed  D  C  Z  .   To  this  end,  define  p  ■  (x,y)  =  min{u-(x,y,z)|z  6  D 
n  C(x,y)}  .   (By  definition  p .  (x,y)  =  oo  if  D  n  C(x,y)  =  0  .)   Like  p.  ,  p  •    is  lower 
semicontinuous.   Hence  P~  (D)  =  {(x,y)|p.(x,y)  =  p.  (x,y)}  is  measurable.   Let  q.:  gr(A) 
-•  Z  be  the  resulting  measurable  selection. 


19 


The  proof  now  divides  into  three  parts.   For  each  x  e  X  ,  let   B(x)  be  the  set  of 
paths  of  game  x  that  are  consistent  with  Nature's  mixed  action  £*(  •  jx)  .   The  first  part  of 
the  proof  shows  that  this  mapping  is  an  upper  semicontinuous  mapping  from  X  into 
J6{9Jl^i  *  Z))  .   The  second  shows  that  the  graph  of  #C  is  closed.   Since  this  graph  is 
clearly  contained  in  that  of  B  ,  all  that  remains  is  to  demonstrate  that   #C  is  non— empty 
valued.   This  is  done  in  the  third  step.   The  second  step  accounts  for  the  bulk  of  the  proof. 

It  is  clear  that  B  is  non-empty  valued,  and  that  its  graph  is  closed.   All  that  we 
need  therefore  show  is  that  it  is  compact  valued.  To  this  end,  let  x  e  X  be  given.   Since 
B(x)  is  closed,  we  need  only  show  that  it  is  tight  in  order  to  show  that  it  is  compact. 
That  is,  we  need  only  show  that,  for  all   e  >  0  ,  there  exists  a  compact   K  c  Y  *  Z  such 

that   A(K)  >  1— e  for  all  A  e  B(x)  .   But  there  certainly  exists  a  compact   K  C  Y*   such 

(V 

that  ffl(K)  >  1— e ,  and  we  may  therefore  let   K  be  the  graph  of  the  restriction  of  C(x,  • ) 


to  K 


ieX* 


AjW 


.  This  completes  the  first  step. 


In  order  to  begin  the  second  step,  suppose  that   (x  ,A  )  6  gr^C]  ,  and  that   (x  ,A  ) 


n'  ny 


n'   n' 


-» (x,A)  .   For  all  i  ,  let   r-   be  player  i 's  payoff  from  A    ,  and  let   it.   be  his  payoff  from 
A  .   From  the  definition  of  ^C(x  )  one  can  deduce  that: 


/Xi(yi)xj(yj)dAn(y,z)  =  [/xi(yi)dAn(y,z)J  |/xj(yj)dAn(y>z) 


.(5.1) 


for  all  i  #  j  and  all  continuous   y.  :  Y.  -»  R  and  y-  :  Y. 
J  Ai       i  AJ       J 


IX0(y0)dAn(y,z)  =  jy0(y0)df0(y0]xn) 


...(5.2) 


for  all  continuous  Xa  '■  Y0  -»  K  ; 


supp(An)cgr(C(xn,.)) 


.(5.3), 


20 


where  C(x  ,  • )  is  regarded  as  a  (possibly  empty— valued)  correspondence  from  Y  to  Z 


n 


(C(xn,y)  is  empty  iff  y  t  A(xn))  ; 

/Xi(yi)ui(x,y,z)dAn(y,z)  =  ^/^(y^dA^z)  ...(5.4) 

for  all  active  i  and  all  continuous  X\  :Y.  -» R ;  and 

rM  >  /pi(xn,y\ani)dAn(y)z)  ...(5.5) 

for  all  active  i  and  all  a  .  e  A.(x  )  . 

ni        p  n' 

Equation  (5.1)  follows  from  the  independence  of  the  marginals  of  A     over  the  Y.  . 
Equation  (5.2)  follows  from  the  fact  that   A     is  consistent  with  Nature's  strategy. 
Relation  (5.3)  follows  from  the  facts  that  the  marginal  \i     of  A     is  supported  on  A(x  )  , 
and  that  its  conditional   u  (•  |y)  is  supported  on  C(x  ,y)  for  all  y  6  A(x  )  .   Equation 
(5.4)  holds  because  the  expectation  of  u-(x  ,-,-)   conditional  on  y.  is   ir  ■   a.s.   This 
follows  from  the  fact  that,  in  equilibrium,  the  probability  mass  of  player  i  's  mixed 
strategy  must  be  confined  to  a  set  of  actions,  each  of  which  yields  him  his  equilibrium 
payoff.   (It  is  important  to  avoid  the  word  'support'  here.   For  player  i  's  continuation 
payoff  /u-(x  ,y,z)di/  (z|y)  need  not  be  continuous  in  y  ,  and  so  the  set  of  pure  strategies 
yielding  him  r.  need  not  be  closed.)  Inequality  (5.5)  states  that  player  i  's  equilibrium 
payoff  must  be  at  least  what  he  would  get  if  he  were  to  deviate  to  a  •  and  if,  following 
this  deviation  and  no  matter  what  the  realised  choices  of  the  other  players,  the  worst 
conceivable  continuation  path  from  his  point  of  view  occurred.   Equation  (5.4)  captures  the 
idea  that  no  redistribution  of  probability  mass  among  equilibrium  actions  is  worthwhile. 
Inequality  (5.5)  captures  the  idea  that  no  redistribution  of  probability  mass  from 
equilibrium  to  out— of— equilibrium  actions  is  worthwhile. 

Now  it  is  easy  to  see  that  relations  (5.1),  (5.2)  and  (5.3)  are  preserved  in  the  limit. 
Also,  since  ic  ■  =  /u-(x  ,y,z)dA  (y,z)  ,   x  .  -» ir.  .  Hence  (5.4)  is  preserved  too.   Lastly,  for 


21 


any  a-  e  A.(x)  ,  there  exist  a  .  e  A-(x  )  such  that  a  .  -»  a-  ,  by  the  lower  semi  continuity 
of  A.  .   Since  also  w-   is  lower  semicontinuous,  (5.5)  too  is  preserved.   The  proof  that   #C 
has  a  closed  graph  will  therefore  be  complete  if  we  can  show  that  the  analogues  of  (5.1)- 
(5.5)  for  x  and  A  ,  which  we  shall  refer  to  as  (5.1')-(5.5'),  imply  that   A  6  ^C(x)  . 

Using  standard  considerations  we  can  deduce  from  (5.1'),  (5.2')  and  (5.3')  that  the 
marginal  /j  of  A  over  Y  is  a  product  measure  supported  on  A(x)  ,  that  the  marginal  of 
/x  over  Y*  is  f*(-jx)  ,  and  that   A  has  an  r.c.p.d.   v  over  Z  such  that   £(-|y)  is 
supported  on  C(x,y)  for  all  y  6  A(x)  .   Now  consider  the  game  with  continuation  paths 
specified  by  v  .  The  marginals  \i-  of  \l  over  the  Y.   need  not  constitute  a  Nash 

equilibrium  for  this  game.   However,  by  (5.4'),  the  set   Y-   of  a- 6  A.(x)   such  that 
deviation  by  active  player  i  from  /x.   to   6     ,  the  Dirac  measure  concentrated  at   a-  ,  will 

1  a-  1 

i 

raise  his  payoff  above  x.   has  //.—measure  zero.  We  may  therefore  alter   v  in  such  a  way 

that  it  remains  an  r.c.p.d.  of  A,  but  such  deviations  are  no  longer  attractive.   We  set 
K- 17)  =  K-  ly)  a  y.  6  Y\Y   for  all  i  ,  and  K-|y)  =  *q.(X|y)  if  7j  e  Yj\Yj  for  all  j 

^  i  but  y.  6  Y-  .  It  is  of  no  importance  how  we  define  v{  •  |  y)  if  y.  e  Y.  for  more  than 
one  i.   (Such  a  y  corresponds  to  a  coordinated  deviation  by  two  or  more  players.)   For 
definiteness,  we  set   ^(-|y)  equal  to  £(-|y)  in  this  case.  Inequality  (5.5')  ensures  that, 
with  the  new  continuation  paths  specified  by  v  ,  deviation  is  no  longer  profitable.   The  //• 
therefore  constitute  a  Nash  equilibrium  with  these  continuations,  and  therefore  A 
€  Jc(x)  . 

Having  proved  that   $C  has  closed  graph,  all  that  remains  is  to  show  that   \?C  is 
non-empty  valued.   To  this  end,  fix  x  and  find,  for  each  i  ,  a  sequence  {y  ■}  that  is 
dense  in  A.(x)  .   For  each  1  <  N  <  od  ,  construct  a  finite  game  with  action  sets   {y  - 1 1  <  n 
<  N}  in  the  natural  way  from  the  existing  game  x  .   And  for  N  =  a>  simply  take  the 
existing  game  x  .  The  family  of  games  obtained  as  N  varies  is  continuous,  and  its 
equilibrium  correspondence  therefore  has  compact  values  and  a  closed  graph.   Moreover 
standard  results  show  that  the  set  of  equilibria  is  non-empty  when  N  <  od  .  It  follows  that 


22 


it  is  non-empty  for  N  =  cd  too.        □ 

Lemma  4.1  captures  the  essence  of  the  backwards  program:  we  begin  with  an  upper 
semicontinuous  correspondence  C  depending  on  the  sequence  of  outcomes  up  to  and 
including  the  current  period,  and  end  up  with  an  upper  semicontinuous  correspondence  'I'C 
depending  on  the  sequence  of  outcomes  preceding  the  current  period.   There  is  however  one 
further  complication.   When  we  come  to  complete  the  proof  of  the  theorem  in  Section  4.4, 
the  continuation  paths  in  Z  will  themselves  be  probability  measures  over  sequences  of 
future  outcomes.   In  the  present,  more  abstract,  setting  this  can  be  captured  by  writing  Z 
=  ^^(W)  ,  where  W  is  itself  a  complete  separable  metric  space.  With  this  notation,  the 
equilibrium  paths  described  by  *C  lie  in    9JC{Y  x  $>JC(W))  .   This  means  that   #C 
cannot  serve  directly  as  input  for  the  next  iteration  of  the  backwards  program.   What  is 
required  instead  is  a  correspondence  ^C  :  X  ->  9JC{Y  x  W)  .   (The  point  is  that  elements 
of   &JC{Y  x  W)  are  probability  measures  over  sequences  of  future  outcomes,  whereas 
those  of   9>Jl[X  x  3>JC{yi))  are  not.)   Such  a  correspondence  can  be  obtained  by  replacing 
elements  of   PJC^&Jt^))  with  their  expectations. 

More  formally,  for  each  xGX  let  ^C(x)  c  ,?1(Y  x  W)  be  the  set  of  measures  A 
such  that: 

(i)  the  marginal  \i  of  A  over  Y  is  a  product  measure  with  supp(/z)  C  A(x)  ; 

(ii)         the  marginal  \l*  of  \i  over  Y*  is  fg(x)  ; 
(iii)         A  possesses  an  r.c. p. d.   v  over  W  such  that   v{  •  |  y)  e  co[C(x,y)]  for  all  y 

e  A(x)  ; 
(iv)        moreover  the  marginals  /j.   of  \i  over  the  Y.   constitute  a  Nash  equilibrium  when 

the  continuation  path  following  outcome  y  is  r/(-  |y)  for  all  y  . 
(Once  again  it  may  be  helpful  to  expand  on  (iv).   Suppose  that  we  are  given  \i- 
6  &JC(k.{xj)  and  a  transition  probability  v  such  that   u(-  |y)  G  co[C(x,y)]   for  all  y 
€  A(x)  .   Then  active  player  i  's  payoff  is   /u-(x,y,^(-  |y))d/i(y)  ,  where  /z  is  the  product 
of  the  (jl-  .   The  //•   constitute  a  Nash  equilibrium  if  no  such  player  can  improve  his  payoff 


23 


by  changing  his  p.  .) 

Lemma  4.2    Suppose  that  the  domain  of  definition  of  u-(x,y,z)  is  convex  in  z  ,  and 
suppose  that  u(x,y,z)  is  linear  in  z  on  its  domain.   Then  *C  is  an  upper 
semi  continuous  mapping  from  X  into    Jtp(?JC(Y  x  W))  . 

It  should  be  noted  that,  when  we  come  to  apply  Lemma  4.2,  u-(x,y,z)  will  be 
defined  as  the  expectation  with  respect  to  z  of  an  underlying  function  of  (x,y,w)  .   So  the 
convexity  of  the  domain  of  u.(x,y,z)  in  z  ,  and  the  linearity  of  u-(x,y,z)  in  z  ,  will  be 
natural  consequences  of  the  problem  structure.   Lemma  4.2  will  be  proved  by  showing, 
roughly  speaking,  that   #C  is  the  image  of  tfC  under  the  projection  that  consists  of 
replacing  elements  of  &J£($>J((W))  by  their  expectations,  and  that  this  projection  is 
continuous. 

Proof    Let   A  G  ^JC{Y  x  Z)  ,  let  \i  be  the  marginal  of  A  over  Y  ,  and  let   v  be  an 
r.c.p.d  of  fi  over  Z.   Let  fj,  =  fi ,  and  let   u{-  |y)  =  /z  d^(z|y)  be  the  expectation  of 
j/(-|y)  for  all  y£Y.   Then  the  projection  X  of  A  onto    M(Y  «  W)  is  the  probability 
measure  with  marginal  \i  and  conditional  v  .   (It  should  be  clear  that  X  is  independent 
of  the  r.c.p.d.  chosen  for  A  .)  The  first  step  of  the  proof  is  to  show  that  projection  is 
continuous. 

To  this  end,  let   {An}  C  9Jt{X  x  Z)  converge  to  A  6  9JC{Y  x  Z)  .   Then,  to  show 
that   {A~n}  C  ?JC(Y  x  W)  converges  to  Xe^(YxW),  it  suffices  to  show  that 

/x(yMw)  dA~n(y,w)  -.  /x(yMw)  dl(y,w) 

for  all  continuous  functions  x-  Y  "♦  K  and  ifr.  W  -» IR  .   Let  t  3>JC(W)  -»  K  be  the 
continuous  linear  functional  defined  by  1(k)  =  /^w)d«(w)  .   Then 


24 


/x(yMw)d!n(y,w)  =  /  Mw)di/n(w|y)  x(y)d**n(y) 


(by  Fubini's  theorem) 


=  / 


K\(-\v))  x(y)d/*n(y) 


(by  definition  of  €} 


=  / 


/<z)di/n(z|y)  x(y)d/xn(y) 


(by  definition  of  the  expectation  of  u  (•  |y)) 


/x(y)^(z)dAn(y)Z) 


(by  Fubini  again,  and  because  //    =  //  ).   But  the  latter  expression  converges  to 
/x(y)4z)dA(y,z)  by  definition  of  weak  convergence,  and  this  integral  is  equal  in  turn  to 
/x(y)V(w)dA"(y,z)  by  reversing  the  above  argument.  This  completes  the  proof  of 
continuity. 

To  complete  the  proof,  it  suffices  to  show  that  gr(^C)  is  the  image  of  gr(^C) 
under  the  projection  that  takes   (x,A)  into  (x,X)  .   Suppose  first  that   (x,A)  G  gr(^C)  . 
Then  the  payoff  from  A  is 


/Uj(x,y,z)dA(y,z)  =  /  /ui(x,y>z)di<z|y)  d/x(y) 


=  /ui(x,y,/zd^(z|y))d/i(y) 


(by  the  linearity  of  u-) 


25 


=  /ui(x,y,K-|y))d/i(y) 

(by  definition  of  V).  But  u{-  |y)  G  co[C(x,y)]  because  supp[*/(-  |y)]  C  C(x,y)  ,  and 
/u-(x,y,P(- |y))d/z(y)  is  by  definition  the  payoff  from  A~  .   So  it  is  easy  to  see  that   (x,A~) 
e  gr(*C)  . 

Suppose  now  that  (x,X)  G  gr(#C)  .  We  need  to  find  (x,A)  G  gr(#C)  of  which 
(x,A~)  is  the  projection.   To  this  end,  let   C(x,y,z)  be  the  set  of  k  G  9Ji{V)  with 
expectation  z  and  such  that  supp[/t]  C  C(x,y)  .  Then  C  has  a  closed  graph,  and 
standard  considerations  show  that   C(x,y)  is  non-empty  iff  z  G  co[C(x,y)]  .   Let 
c:  gr(co[C])  -»  &JC(7j)  be  a  measurable  selection  from  the  restriction  of  C  to  gr(co[C])  , 
let   u{  ■  |  y)  =  c(x,y,F(  ■  |  y))  for  all  y  G  A(x)  ,  and  let   v{  •  |  y)   be  arbitrary  for  y  £  A(x) 
(for  example,  one  could  set   v{-  |y)  =  &-,   ,   x  for  such  y).   Then  it  can  be  checked  as 
above,  using  the  linearity  of  u-  ,  that  the  measure  A  obtained  by  combining  the  marginal 
Jl  with  the  transition  probability  v  lies  in   ^C(x)  .        □ 

Lemma  4.2  is  the  basic  result  justifying  the  backwards  program.   We  conclude  the 
present  subsection  by  relating  it  to  the  work  of  Simon  and  Zame  (1990). 

The  standard  existence  theorem  for  normal— form  games  can  be  expressed  as 
follows.   If  each  player's  strategy  set  is  a  compact  metric  space,  and  if  there  is  a  continuous 
function  associating  a  vector  of  payoffs  with  each  vector  of  strategies,  then  there  exists  a 
vector  of  (possibly  mixed)  strategies  that  forms  a  Nash  equilibrium.   Simon  and  Zame 
(1990)  extended  this  result  by  showing  that  if,  instead,  there  is  an  upper  semi  continuous 
correspondence  associating  a  convex  set  of  payoffs  with  each  vector  of  strategies,  then  there 
exists  a  selection  from  this  correspondence  and  a  vector  of  (possibly  mixed)  strategies  that 
forms  a  Nash  equilibrium  when  the  payoff  function  is  given  by  this  selection. 

In  order  to  explain  how  Lemma  4.2  extends  the  result  of  Simon  and  Zame,  it  is 
helpful  to  reinterpret  their  work  slightly.   Much  as  we  do  in  the  proof  of  Lemma  4.1,  they 
consider  a  sequence  of  finite  approximations  to  their  basic  game.   Suppose  these 


26 


approximations  are  indexed  by  N  ,  suppose  that  the  mixed  strategies  /a^-   constitute  a 
Nash  equilibrium  of  game  N  for  some  selection  from  its  payoff  correspondence,  and  let  the 
equilibrium  payoffs  of  this  equilibrium  be  7iYr-  .  Then  their  proof  shows  that,  if  the  mixed 
strategies  /a-   and  payoffs   ir-   constitute  a  limit  point  of  the  strategies  /a^-   and  payoffs 
7iVr-  ,  then  there  exists  a  selection  from  the  payoff  correspondence  of  the  original  game  such 
that  the  \i-   constitute  a  Nash  equilibrium  with  equilibrium  payoffs   ir-   when  the  payoff 
function  is  given  by  this  selection.  In  other  words,  they  demonstrate  a  limited  form  of 
upper  semicontinuity  for  a  particular  continuous  family  of  games. 

Lemma  4.2  therefore  extends  the  result  of  Simon  and  Zame  in  two  ways.12  The 
first  is  relatively  minor:  it  shows  that  upper  semicontinuity  holds  for  a  general  continuous 
family  of  games.   The  second  is  more  substantive.  The  result  of  Simon  and  Zame  tells  us 
about  the  behaviour  of  the  equilibrium  strategies  and  equilibrium  payoffs,  but  it  tells  us 
very  little  about  the  selections  from  the  payoff  correspondence  that  enforce  these  strategies. 
Lemma  4.2,  by  contrast,  concerns  a  comprehensive  description  of  equilibrium.   Such  a 
description  of  equilibrium  is  not  without  intrinsic  interest.   But  its  main  significance  for 
the  purposes  of  the  present  paper  is  that  it  allows  us  to  avoid  a  measurable  selection 
problem  that  would  otherwise  arise  when  we  come  to  program  forwards.13 


12  As  it  stands  Lemma  4.2  is  not  a  direct  generalisation  of  the  result  of  Simon  and 
Zame.  But  it  can  easily  be  adapted  to  obtain  one.  To  do  so,  view  Z  ,  not  as  a  space  of 
probability  measures  over  a  compact  metric  space,  but  more  generally  as  a  compact 
metrisable  subset  of  a  locally  convex  topological  vector  space.   (This  perspective  is  more 

general,  and  includes  the  possibility  that   Z  is  a  compact  subset  of  IR   .  If  Z  is  a  compact 

subset  of  IR    ,  then  one  can  define  u-(x,y,z)  =  z.  .)   And  define   ^C(x)  in  a  more  limited 

way,  to  consist  of  pairs  (/a,7t)  G  &J((Y)  *  IR   ,  where  /a  is  the  distribution  of  a  Nash 
equilibrium  and  ir  is  its  equilibrium  payoff  vector.   (This  more  limited  definition  of  $C  is 
the  price  one  must  pay  for  the  more  general  Z  .)  Then  a  proof  almost  identical  to  that  of 

Lemma  4.2  shows  that  $C  is  a  continuous  projection  of  $C,  and  therefore  upper 
semi  continuous. 

13  It  would  be  tempting  to  conduct  the  backwards  program  as  follows.   Let   II:  gr(A) 

-» IR    be  an  upper— hemicontinuous  convex— valued  correspondence  describing  possible 

continuation  payoffs.   Then  $11:  X  -» IR   ,  the  correspondence  describing  possible  Nash 
equilibrium  payoffs  for  the  current  period,  is  upper  hemicontinuous;  and  so  therefore  is 
co^II]  ,  the  continuation— payoff  correspondence  for  the  next  iteration  of  the  backwards 
program.  This  simple  program  is,  however,  difficult  to  reverse.   For,  in  order  to  reverse  it, 


27 


4.3    The  Forwards  Program 

In  this  subsection  we  develop  the  analysis  necessary  for  the  forwards  program.   For 
this  purpose  we  shall  need  a  measurable  selection  c  from  #C  .  This  function  can  be 
thought  of  as  part  of  the  output  of  previous  iterations  of  the  forwards  program.  It 
describes  the  equilibrium  paths  that  must  be  followed  from  the  current  period  onwards  if 
the  strategies  calculated  for  earlier  periods  by  the  previous  iterations  of  the  forwards 
program  are  really  to  constitute  part  of  an  equilibrium. 

Lemma  4.3    There  exist  measurable  functions  A:  X  x  [0,1]  -»  ^J({Y  *  W)  and  w.X  *  Y 

x  [0,1]  -»  $>JC{W)  such  that,  for  all  (x,£)  e  X  *  [0,1]  : 

(i)  the  marginal  /x(-|x,£)  of  A(-|x,£)  over  Y  is  a  product  measure  such  that 

supp[/x(-|x,0]  C  A(x)  ; 
(ii)         themarginal^0(-|x,£)of/z(-|x,£)  over  Y0  is  ity{-\x); 
(iii)         i/(-|x,-,£)  is  an  r.c.p.d.  of  A(«|x,£)  over  W  such  that   u{-  |x,y,£)  €  co[C(x,y)] 

for  all  y  6  A(x); 
(iv)         the  marginals  /j.(-|x,£)  of  /x(»|x,£)  over  the  Y-   constitute  a  Nash  equilibrium 

when  the  continuation  path  following  outcome  y  is  v{-  |x,y,£)  for  all  y  e  A(x). 
Moreover  /A(-  |x,£)d£  =  c(x)  for  all  x  e  X  . 

Lemma  4.3  captures  the  essence  of  the  forwards  program.   It  requires  some 
explanation.   First  of  all,  conditions  (i)— (iv)  imply  that   A(-  |x,£)  €  ^C(x)  .   The  condition 
/A(-  |x,£)d£  =  c(x)  therefore  tells  us  that,  if  the  equilibrium  path  is  chosen  from   ^C(x) 
on  the  basis  of  the  current  public  signal   £  e  [0,1]  ,  then  the  overall  path  will  be  precisely 


one  must  be  able  to  find,  for  any  measurable  selection  ir  from  #11  ,  a  measurable 
selection  $7r  from  II  such  that  the  game  with  payoff  function  $7r(x,-)  has  a  Nash 
equilibrium  with  payoff  7r(x)  .   This  is  not  a  standard  measurable— selection  problem.   For 
what  is  actually  required  is  that  we  select,  for  each  x  ,  a  function  $7t(x,-)  ,  and  that  these 
choices  of  functions  be  coordinated  in  x  to  obtain  a  single  measurable  function  3>7r .   (A 
version  of  this  non— standard  problem  did  arise  for  Mertens  and  Parthasarathy  (1987; 
Lemma  1  of  Section  6,  pp  41— 42).  In  their  case,  however,  the  functions  $7r(x,-j   could  be 
taken  to  be  continuous.) 


28 


c(x)  .   (If  Lebesgue  measure  is  the  marginal  distribution  over  [0,1]  ,  and  if  A(-  |x,-)  is  the 
conditional  distribution  over  Y  x  W  ,  then  /A(-|x,^)d^  is  the  marginal  distribution  over 
Y  x  W  .)    Secondly,  the  lemma  goes  beyond  the  simple  assertion  that  there  exists   A:  X 
x  [0,1]  -»  ?JC{Y  x  W)  such  that  A(-  |x,£)  G  *C(x)  for  all  (x,£)  .   For,  while  this 
assertion  would  automatically  imply  the  existence  of  a  v  that  was  measurable  in  y  for 
each  (x,£)  ,  the  proposition  delivers  a  v  that  is  measurable  in  (x,y,£)  jointly.  Thirdly, 
the  \i-\  X  x  [0,1]  -•  &J((Y.)  are  precisely  the  strategies  of  the  players  for  the  current 
period.   Fourthly,  the  restriction  $c  of  v  to  gr(A)  x  [0,1]  describes  the  equilibrium 
paths  that  must  be  followed  from  the  next  period  onwards.  It  serves  as  the  input  for  the 
next  iteration  of  the  forwards  program. 

Proof    Most  of  the  considerations  necessary  for  the  proof  have  already  arisen  in  the  proof, 
of  Lemmas  4.1  and  4.2.   The  present  proof  will  therefore  be  brief. 

The  first  step  is  to  find  a  measurable  d:  X  -•  9JL ( $>JC(Y  *  W))  such  that,  for  all 
x  G  X  ,  supp(d(x))  c  \PC(x)  and  the  expectation  of  d(x)  is  c(x)  .   This  can  be  done  using 
an  argument  very  similar  to  one  used  in  the  proof  of  Lemma  4.2.   Next,  we  need  a 
measurable  mapping  A:  X  x  [0,1]  -»  &J£(Y  x  W)  such  that,  when  [0,1]  is  given  Lebesgue 
measure,  the  distribution  of  the  random  variable  A(-  |x,«):  [0,1]  -»  9Jl{Y  x  YV)  over 
^Ji(Y  x  W)  is  precisely  d(x)  for  all  x  .   Such  a  mapping  can  be  obtained  by  applying  a 
'measurable'  representation  theorem,  e.g.  Gihman  and  Skorohod  (1979;  Lemma  1.2,  p  9). 
The  third  step  is  to  find  a  measurable  mapping  u:  X  x  Y  x  [0,1]  -»  &Jt(W)  such  that 
£(-|x,-,£)  is  an  r.c.p.d.  of  A(- |x,£)  for  all  (x,£)  e  X  x  [0,1]  .   In  other  words,  we  need  a 
'measurable'  version  of  the  standard  result  which  asserts  the  existence,  for  each  (x,£)  ,  of 
an  r.c.p.d.   v{-  |x,-,£)  of  A(-  |x,£)  .   Since  this  is,  in  effect,  the  central  step  of  the 
forwards  program,  and  since  we  have  not  been  able  to  find  precisely  the  result  we  need  in 
the  literature,  we  sketch  a  proof  of  this. 

Let  /x(-  |x,£)  be  the  marginal  of  A(«  |x,£)  over  Y  .   Let   {B    1 1  <  n  <  <d}  be  a 
sequence  of  sets  that  generates  the  Borel  a— algebra  on  Y  .   For  each  N  ,  let    5?™  be  the 


29 


a-algebra  generated  by  {B   1 1  <  n  <  N}  .  And  let  v  be  a  fixed  element  of   ^l#(W)  . 
Since   &■**  is  finite  we  can  give  an  explicit  formula  for  an  r.c.p.d.  of  A(-|x,£) 

given  the  a— algebra   5?-^  on  Y  .   For  each  y  G  Y  and  each  measurable  W  C  W  we  set 


,,;,        -      A(YxW|x,Q 
"N(W|x,y,£)  =  -L- - — l-x^L 

M(Y|x,0 


if  Y  is  the  atom  of   &N  containing  y  and  /i(Y|x,£)  >  0  ,  and 


N 


z,N(W|x,y,£)  =  KW) 

otherwise.   It  is  immediate  from  these  formulae  that   u-^(-  |x,y,£)  is  jointly  measurable  in 
(x,y,£)  ,  and  therefore  that  the  convergence  set  K  of  u-^  is  measurable.   Set   v 
=  lim!/v,on  K  and  v  —  v  outside  K  . 

Certainly  u  is  measurable.   Also,  since  i/^(-|x,-,£)  is  effectively  a  version  of  the 
conditional  expectation  of  A(-  |x,£)  given  &„  ,  the  martingale  convergence  theorem 
implies  that    1  im  i/M(-  |x,-,^)  exists  //(•  |x,£)  —  a.s.,  and  that  u(-\x,',£)  is  a  version  of 

what  is  effectively  the  conditional  expectation  of  A(-  |x,£)  given  the  Borel  a— algebra  on 
Y.  That  is,  u{-  |x, •,£)  is  an  r.c.p.d.  of  A(-|x,£).   This  completes  the  third  step. 
There  are  two  difficulties  with  u  as  it  stands:  it  may  not  be  the  case  that 
u(-  |x,-,£)  e  co[C(x,y)]  for  all  y  6  A(x)  ;  and  the  marginals  /z(-|x,£)  of  A(-|x,£)  need 
not  constitute  a  Nash  equilibrium  when  the  continuation  path  following  outcome  y  is 
u(-  |x,y,£)  .  The  fourth  and  final  step  constructs  a  v  that  does  have  these  properties  from 
v  .  To  obtain  the  first  property,  let  q  be  any  measurable  selection  from  co[C]  ,  and 
redefine  i>(-|x,y,£)  to  be  q(x,y)  wherever  it  does  not  lie  in  co[C(x,y)]  for  some  y 
G  A(x)  .   To  obtain  the  second:  for  each  i  ,  let  q.   be  the  function  defined  in  the  proof  of 

Lemma  4.1;  let   7r.(x,£)  =  Ju-(x,y,v(-  |x,y,^))d/i(y|x,^)  ;  and  let  Y.(x,£)  be  the  set  of  a- 


30 


6  A-(x)   such  that   /u.(x,y\a.,^(-  |x,y,£))d/z(y|x,£)  >  ^(x,^)  .  Then  we  may  set 

v{-  |x,y;0  =  qi(x,y)  if  7.  6  Y^YjCx.O  for  all  j  *  i  but  yj  £  Y(x,0  ,  and  v{-  |x,y,£) 
=  £(•  |x,y,£j  otherwise.   Standard  considerations  show  that  the  resulting  v  is  measurable, 
and  the  fact  that   A(-  |x,£)  e  C(x)  implies  that  u(-  |x,»,£)  is  still  an  r.c.p.d.  of 

A(-|x,£).        o 

4.4  Completion  of  the  proof 

Before  proceeding,  we  need  to  define  the  equilibrium  correspondence  for  a  family  of 
games.   In  making  this  definition  we  encounter  a  minor  technicality.  We  have  defined  an 
equilibrium  as  a  strategy  profile  that  is  in  Nash  equilibrium  in  every  subgame.   But  a 
strategy  profile  specifies  players'  behaviour  for  the  entire  family  of  games  that  they  might 
face.   Hence  an  equilibrium  will  generate  a  whole  family  of  paths,  one  for  each  game  x 
6  X  .   We  therefore  define  E..  (x  )  ,  the  set  of  equilibrium  paths  of  game  x    ,  to  be  the  set 
of  paths   A  such  that   A  is  the  equilibrium  path  generated  in  game  x     by  some 

equilibrium.   Similarly,  reinterpreting  our  basic  family  of  games  as  a  family  parameterised 

t— 1        t— 1 
by  x       e  X        ,  we  can  define,  correspondences  E.   for  t  >  2  . 

Theorem  4.1    Suppose  that  (Al)— (A3)  hold.  Then  E,    is  an  upper  semi  continuous  map 
from  Xt_1  into    J6{9Jt{**_ .  Yj). 

The  main  interest  of  Theorem  4.1  derives  from  the  fact  that  it  implies  existence 
(see  Section  5).   However  upper  semicontinuity  is  itself  of  interest:  it  is  reassuring  to  know 
that  our  concept  of  equilibrium  possesses  this  standard  regularity  property. 


Proof    The  proof  proceeds  in  three  steps.  Let  B.(x  —  )  be  the  set  of  paths  beginning  in 
period  t  that  are  consistent  with  Nature's  strategy.   The  first  part  of  the  proof  shows  that 
B.    is  an  upper  semicontinuous  map  from  X  —     into    JGi^JC^  ®Y  )).   In  the  second,  a 

I  S — X     s 

sequence  of  correspondences  {C.  1 1  <  t  <  cd}  is  constructed,  each  of  which  is  upper 


31 


semicontinuous.   In  the  third  it  is  shown  that   E.  =  co[C.]  for  all  t  . 

For  notational  convenience,  we  show  that  B1   is  upper  semicontinuous.  To  this 
end,  let  B?(x  )  be  the  set  of  marginals  over  «     ,Y    of  measures  in  B,(x  ).   Then  a 
backwards  programming  argument  similar  to,  but  simpler  than,  that  developed  in  Lemmas 
4.1  and  4.2  shows  that   B,  is  an  upper  semicontinuous  map  from  X     into 
J6{9Jt{*  _,  Y  )).   But  a  sequence  of  measures  {A  }  C  ^("/iY.)  converges  iff  its 
marginals  over   PJCi*  _.Y  )  converge  for  all  t.   So  B,  ,  too,  is  upper  semicontinuous. 

This  completes  the  first  step. 

T 

Turning  to  the  second  step,  let  T  >  1  be  given.  For  each  t  >  T  ,  let  C.  =  B.  . 

T  T 

And,  for  each  1  <  t  <  T  ,  let  C  .    be  defined  inductively  by  the  formula  C  ,  = 

T  T    t— 1 

$,  (C .  , , )  .   (In  other  words,   C  .  (x      )  is  the  set  of  equilibrium  paths  that  are  obtained 

T 
for  period  t  onwards  when  the  continuation  paths  can  be  chosen  from  co[C,     ,]  .)  Then 

T 
the  backwards  program  ensures  that  each  C,    is  upper  semicontinuous. 

T  T         S 

Now  define  C.  =  nt™,C.    for  each  t  >  1.  It  is  obvious  that   C,  cC.  whenever  T 

>  S.   Hence  C,  too  is  upper  semicontinuous.  It  is  also  the  case  that   ^t(C,  ,  ,)  C  C..  To 

see  this,  let   A  6  [^t(C.  ,  1)](x      )  .   By  definition,   A  can  be  enforced  by  continuation 

T 

paths  from  C,    ,  .   But  paths  in  C.  ,,    belong  a  fortiori  to  C,  ,-,  .   Hence 

A  6  [^t(C^+1)](xt_1)  for  all  T  .   Since  Ct  =  nT=i*t(C T+l^  ' the  rec*uired  conclusion 

T 
follows.   Finally,   C.  C  *+(C.  , ,)  .   To  see  this  note  that  the  correspondences   C  .  ,  ,  for 

1  <  T  <  od  together  with  C,,,    for  T  =  cd  form  a  continuous  family  of  games  with 

parameters  x         and  T  .   And  Lemma  4.2  shows  that  the  equilibrium  correspondence  of 

such  a  family  is  upper  semicontinuous.  It  follows  that  any  A  6  C.  (x  —  )  ,  which  is 

certainly  a  limit  point  of  points  in   [^.(C.     ,)](x      ),  also  lies  in  [*t(C,  ,  ,)](x      ). 

Overall,  then,  we  obtain  a  sequence  of  upper  semicontinuous  correspondences 
{C.|  1  <  t  <  a) }  such  that   C,  =  ^t(C.  ,  ,)  for  all  t  .   It  remains  only  to  show  that  E, 
=  co[C,]  for  all  t  .  We  do  this  for  t  =  1  only.   The  other  cases  are  analogous. 

We  show  first  that  co[CJ  C  E,  .  It  suffices  to  show  that  any  measurable  selection 
c,   from  co[CJ  is  also  a  selection  from  E1  .   So  let  C.     be  any  such  selection.   Then 


32 


repeated  application  of  the  forwards  program  starting  with  c,    yields,  for  all  t  >  1  , 
strategies  f,.   and  a  measurable  selection  c.,,   from  co[C.,J  such  that:  (i)  for  all 
(x      ,()e  H~    ,  the  f.-(x  —  ,£  )  constitute  a  Nash  equilibrium  in  period  t  when  the 
continuation  paths  are  given  by  c,  ,  ,(x      ,•>£  )  and  (")  the  paths  obtained  from  period  t 
onward  when  the  f..   are  employed  in  period  t  and  the  continuation  paths  are  given  by 
c.,,    are  precisely  those  required  by  c.  .  It  follows  from  (i)  and  a  standard  dynamic 
programming  argument  that  the  strategy  profile  f  constitutes  an  equilibrium.   And  it 
follows  from  (ii)  that  the  equilibrium  path  in  game  x     is  precisely  c,  (x  )  .  This 
completes  the  proof  that  co[CJ  c  E,. 

We  now  show  that  E,  C  co[CJ  .   To  this  end,  let  f  =  <f..  |t  >  1,  i  €  X<>   be 
any  subgame-perfect  equilibrium  in  extended  strategies.   For  each  t  >  1  ,  let  c ,:  H  ~~ 
-»  3>JC(X  *  Y  )  describe  the  paths  followed  from  period  t  onwards  when  the  strategy 

profile  f  is  employed.   The  mapping  c,    inherits  measurability  from  f .   Let  T  >  1  be 

T 
given.   Now  certainly  c™,,   is  a  selection  from  co[Cy,,].   Also,  suppose  that   c,,1    is  a 

T 

selection  from  co[C.     J  for  some   1  <  t  <  T.   Since  f  is  an  equilibrium,  the  f,-   specify 

Nash  equilibria  for  period  t  when  the  continuation  paths  are  given  by  c.    ,  .  It  follows 

T 
at  once  that  c,   is  a  selection  from  co[C  ,  ]  .  Proceeding  inductively  we  therefore  conclude 

T 
that  c,   is  a  selection  from  co[C,  ]  .  Moreover,  allowing  T  to  vary,  we  obtain  that  C,   is 

a  selection  from  co[CJ  .   This  completes  the  proof.        □ 


33 


5    Existence  with  and  without  Public  Signals 

This  section  provides  a  systematic  discussion  of  the  question  of  existence  of 
equilibrium  in  the  dynamic  model  introduced  in  Section  3.   It  begins  with  a  recapitulation 
of  the  main  result  for  this  case,  namely  that,  when  public  signals  are  allowed,  any  family  of 
games  satisfying  the  standing  assumptions   (Al)— (A3)  possesses  an  equilibrium.   However, 
the  section  also  addresses  two  further  questions.   Under  what  circumstances  can  existence 
be  obtained  without  the  use  of  public  signals?  And  is  the  extension  of  the  concept  of 
subgame— perfect  equilibrium  obtained  by  allowing  the  observation  of  public  signals 
sufficiently  rich? 

5.1        Existence  with  Public  Signals 

The  following  result  is  an  immediate  corollary  of  Theorem  4.1: 

Theorem  5.1    Suppose  that  (Al)— (A3)  hold.   Then  there  exists  a  subgame-perfect 
equilibrium  in  extended  strategies. 


Proof    Pick  any  x    eX    .   By  Theorem  4.1,  E,(x  )  is  non-empty.   Pick  any  A  6  E.(x  ). 
Then,  by  definition  of  E..(x  )  ,  there  exists  an  equilibrium  of  the  family  of  games  that 
generates  equilibrium  path  A  in  game  x    .  In  particular,  there  exists  an  equilibrium! 


The  potential  drawback  of  this  result  is  that,  in  order  to  obtain  the  existence 
equilibrium  in  what  is  intended  to  be  a  purely  non-cooperative  situation,  it  introduces  an 
extraneous  element  into  the  model,  namely  public  signals.  It  is  therefore  of  some  interest 
to  discover  circumstances  under  which  equilibrium  exists  even  without  this  element. 


34 


5.2  Existence  without  public  signals 

There  are  at  least  three  classes  of  game  in  which  public  signals  are  not  needed  for 
existence:  games  of  perfect  information,  finite— action  games,  and  zero— sum  games.   The 
results  for  such  games  are  summarised  in  Theorems  5.2,  5.3,  and  5.4  respectively. 

Theorem  5.2    Suppose  that  (Al)— (A3)  hold,  and  that  the  family  of  games  has  perfect 
information.  Then  a  subgame— perfect  equilibrium  exists.        a 

Proof    Oddly  enough,  Theorem  5.2  is  not  an  immediate  corollary  of  Theorem  5.1. 
(Intuitively  it  would  seem  obvious  that,  when  only  one  player  acts  at  a  time,  public  signals 
do  not  allow  players  to  achieve  anything  they  could  not  achieve  using  mixed  strategies. 
But  there  is  a  minor  complication:  players  observe  not  only  the  current  signal,  but  also 
past  signals.)   However,  an  equilibrium  can  be  constructed  using  the  correspondences   C, 
used  in  the  proof  of  Theorem  4.1.   We  begin  with  any  measurable  selection  c,    from  C-.  , 
and  define  f,.(x  )   to  be  the  marginal  of  c,(x  )  over  Y...  .   Next,  let  c«  be  a  measurable 
selection  from  co[C  J   that  enforces  c,  .   Since  the  game  is  a  game  of  perfect  information, 
co[C2]  =  C^  •   Hence  c„  is  also  a  measurable  selection  from  C«  .   So  we  may  define 
L.(x  )  to  be  the  marginal  of  c2(x  )  over  Y,.  .   And  so  on.        o 

Indeed,  one  can  even  show  that  there  exists  a  subgame-perfect  equilibrium  in 
which  players  employ  pure  actions  only.   We  do  not  demonstrate  this  here,  since  the 
argument  does  not  appear  to  be  a  simple  corollary  of  the  construction  that  we  have 
employed.14  Note  that  the  proof  of  Theorem  5.2  shows  incidentally  that  allowing  public 
signals  in  a  game  of  perfect  information  does  not  enlarge  the  set  of  equilibrium  paths. 


14         See  Harris  (1985)  for  a  treatment  of  the  case  of  perfect  information.   The  argument 
used  there  can  easily  be  adapted  to  the  present  context.   Alternatively,  one  can  work 
directly  with  the  apparatus  used  here,  taking  care  to  purify  the  selections  as  one  programs 
forward. 


35 


Let  us  say  that  a  family  of  games  is  of  finite  action  if  X     is  finite,  and  if,  for  all 


t  >  1  and  all  x"   1  6  X1   *  ,  A^x1   *)  is  a  finite  set. 


Theorem  5.3    Suppose  that  (Al)— (A3)  hold,  and  that  the  family  of  games  is  of  finite 
action.   Then  a  subgame— perfect  equilibrium  exists.   Moreover,  for  each  x    6  X    ,  the  set 
of  equilibrium  paths  is  compact. 


Proof    Theorem  5.3  is  more  or  less  well  known,  and  there  is  more  than  one  way  of  proving 
it.   (See  Fudenberg  and  Levine  (1983)  for  a  different  proof.)   So  we  merely  note  that  it  can 
be  proved  by  a  simple  adaptation  of  the  methods  of  this  paper.   The  basic  idea  can  be 
expressed  in  terms  of  the  notation  of  Section  4.2  as  follows.  Instead  of  defining  #C(x)  to 
be  the  set  of  equilibrium  paths  that  can  be  obtained  by  choosing  continuation  paths  from 
co[C]  ,  one  defines  it  to  be  the  set  of  equilibrium  paths  that  can  be  obtained  by  choosing 
continuation  paths  from  C.   Because  X  is  finite,  and  because  the  A.(x)  are  finite  for  all 
x  e  X  ,  this  new  definition  still  yields  an  upper  semicontinuous  correspondence. 

It  should  be  noted  that  both  finiteness  assumptions  are  needed.  Indeed,  suppose 
that  X  is  finite  but  that  one  or  more  of  the  A-(x)   are  infinite  for  some  x.   Then  it  can 
happen  that  there  is  a  sequence  of  equilibria  of  game  x,  each  of  which  involves  a  public 
signal  generated  endogenously  by  players'  randomisation  over  their  action  sets  A.(x),  and 
that  this  sequence  converges  to  a  limiting  equilibrium  which  does  not  generate  any  such 
signal.   Suppose,  on  the  other  hand,  that  the  A.(x)  are  finite  for  all  xgX  but  that  X  is 
infinite.  Then  it  can  happen  that  there  is  a  convergent  sequence  of  games  in  X  ,  and  that 
there  is  at  least  one  player  for  whom  at  least  one  player  for  whom  at  least  two  actions 
coalesce  in  the  limit.   If  this  player  randomises  over  his  two  actions,  then  he  generates  a 
public  signal  which  disappears  in  the  limit. 

It  is  precisely  in  order  to  compensate  for  the  potential  disappearance  of  such 
endogenously  generated  public  signals  that  the  exogenous  public  signals  are  used  in  our 
basic  existence  theorem.        o 


36 


To  complete  our  discussion  of  the  existence  of  subgame— perfect  equilibrium,  we 
consider  the  case  of  a  family  of  zero— sum  games.   (A  family  of  games  is  of  zero  sum  if   J<A 
contains  precisely  two  elements,  and  if  E.    j/^\  =  0  .) 

Theorem  5.4    Suppose  that  (Al)— (A3)  hold,  and  that  the  family  of  games  is  of  zero  sum. 
Then  a  subgame— perfect  equilibrium  exists. 

In  order  to  understand  this  result,  it  is  helpful  to  recall  some  facts  about  one- 
period  zero— sum  games.   It  is  well  known  that  each  players'  set  of  optimal  strategies  is 
convex  in  such  a  game.   But  the  set  of  equilibria,  regarded  as  the  set  of  probability 
distributions  over  outcomes,  need  not  be.   Hence,  even  in  a  zero— sum  game,  the  set  of 
equilibria  will  be  enlarged  if  players  are  allowed  to  observe  a  public  signal.   The  set  of 
equilibrium  payoffs,  on  the  other  hand,  will  not.   For  the  public  signal  merely  allows  the 
two  players  to  co-ordinate  on  a  choice  of  Nash  equilibrium  for  the  game,  and  all  Nash 
equilibria  have  the  same  equilibrium  payoff  (namely  (v,  — v)  ,  where  v  is  the  value  of  the 
game). 

Proof    Note  first  that  our  proof  of  the  existence  of  equilibrium  shows,  in  particular,  that 
the  extended  family  of  games  has  a  Nash  equilibrium.   It  follows  immediately  from 
standard  considerations  that  this  family  also  has  a  value.  That  is,  there  exists  a 
measurable  mapping  v,  :  X    ->  [R  such  that   v,  (x  )  is  the  value  of  extended  game  x    . 
(Indeed,  since  v.,(x  )  is  simply  player  l's  payoff  from  any  path  in  E,(x  ),  and  since  E,  is 
upper  semicontinuous,  v,    is  actually  continuous.   We  do  not  need  this  fact,  however.) 
Similarly,  there  exist  mappings  v  ■  X  ~~   -» IR  such  that  v.(x      )  is  the  value  of  the 
extended  game  beginning  in  period  t. 

With  these  considerations  in  mind,  we  can  construct  a  subgame— perfect 
equilibrium  of  the  original  family  of  games  as  follows.   The  construction  follows  the 
construction  of  an  equilibrium  of  the  extended  game,  but  takes  care  to  avoid  the  need  for  a 


37 


public  signal.   The  first  step  is  to  find  a  measurable  selection   c,   from  C,  .   Since  c,    is  a 
measurable  selection  from  C,    rather  than  co[C,]  ,  its  marginals  over  Y,    are  already 
product  measures.   So  we  may  define  strategies  f,.:  X    -»  ^^(Y.-)  by  setting  f,.(x  )  to 
be  the  marginal  of  c,(x  )  over  Y...  .   Similarly,   c,    generates  a  measurable  selection  Cr, 
from  co[CJ  that  enforces  the  f,.  .   But  all  A  e  co[C2(x  )]   generate  the  same 
continuation  payoff,  namely  v2(x  )  .  In  particular,  if  c«  is  a  measurable  selection  from 
C9  then  (L  generates  the  same  continuation  payoffs  as  c«  •  Hence  the  f,  •  can  be 
enforced  by  <L  just  as  well  as  by  c«  •   (The  continuation  paths  specified  by  c«  do  not 
necessarily  constitute  an  r.c.p.d.  of  c,  ,  of  course.)    Iterating  this  argument  we  obtain 
strategies  £.:  Xt_1  -»  ?JC{Y.^  for  all  t  >  2  as  well,    o 

5.3    Approximation  by  finite— action  games 

Suppose  that  we  are  given  a  single  game  (i.e.  a  family  of  games  for  which  X     is  a 
singleton)  satisfying  (Al)— (A3).   Then  a  finite— action  approximation  to  this  game  is, 
roughly  speaking,  a  new  family  of  games  indexed  by  n  e  IN  U{od}   such  that:  the  game 
corresponding  to  n  =  oo  is  the  original  game;  for  each  n  6  IN  ,  the  game  corresponding  to  n 
is  of  finite  action;  and  taken  as  a  whole  the  new  family  of  games  satisfies  (Al)-(  A3).   (Here 
(Al)  ensures  that  players'  action  sets  in  the  original  game  are  continuously  approximated; 
(A2)  ensures  that  Nature's  behaviour  is  continuously  approximated;  and  (A3)  ensures  that 
active  players'  payoffs  are  continuously  approximated.) 

Finite-action  approximations  to  a  single  game  satisfying  (Al)-(A3)  certainly  exist. 
Indeed,  we  can  even  construct  a  finite-action  approximation  to  a  family  of  games  satisfying 
(Al)-(A3).   To  do  this,  we  need  the  following  ingredients.   First,  for  each  t  >  1   and  each 
i  6  J,  let   {yjj|n  6  IN}  be  dense  in  Yt-  ;  and  let  {yj|n  6  IN}  be  dense  in  YQ  =  X°. 
Secondly,  for  each  n  €  IN  ,  let  Y||  =  {ym|  1  <  m  <  n}  .   Thirdly,  for  each  t  >  1,  i  G  J  and 
n  6  IN,  let   o£j:X  —   -»  Y..   be  a  continuous  function  such  that,  for  all  x  —    e  X  —  : 
a..(x  —  )  belongs  to  A,.(x  —  )  ,   and   ay(x  —  )  is  as  close  to  y?.   as  any  other  point  in 
A..(x      ).   (Such  a  function  exists  because  A,.:X  —   -»  J£(Y..)  is  continuous.  In 


38 


n    vt-l 


Nature's  case  the  function  is  simply  constant  with  value  y.*  .)   And  let   a.-:  X 
-.  MYti)  be  defined  by  ^(x1-1)  =  {aj^x*-1)!  1  <  m  <  n}  for  all  xt_1  e  Xt_1. 
Fourthly,  for  each  t  >  1  and  n  6  IN,  let   ^\*:  X  —   -»  &JC(Y.q)  be  a  continuous  function 
such  that,  for  all  x       e  X      :  supp  [</>t(j(x      )]  C  {y  .  A\  <  m  <  n};  and  <Ptrt(x      )  is  as 
close  to  ffrt(x      )  as  another  measure  in    &J((Y.  *)  whose  support  is  contained  in 
{y™0  1 1  <  m  <  n}  .   Finally:  let  Yj  =  XQ  ;  for  all  t  >  1  ,  i  6  Jj6  and  xt_1  6  Xt_1  ,  let 
^.(x1-1)  =  Ati(xt_1)  ;  and  for  all  t  >  1  and  xt_1  e  Xt_1  let   ^0(xt_1)  =  ft0(xt_1)  . 
In  terms  of  these  ingredients  we  can  define  a  larger  family  of  games,  which  can  be 
interpreted  as  a  sequence  of  families  of  finite-action  games  approximating  the  original 

ssr\ 

family,  as  follows.   First,  let  X     be  the  union  over  n  6  IN  U  {qd}  of  the  sets   {n}  *  Yn  . 

^t— 1  ^ 

Secondly,  suppose  that  X        has  been  defined  for  some  t  >  1  .  Then:  for  all  ie  7,   A,. 

S\i       1  s\  s\  /\  /\ 

is  the  restriction  of  aj-(-)  to  X  —  ;  A.  =  x-  A  •  and  X.  =  gr(A.)  .  Thirdly,  for  all 
t  >  1  ,  f.0  is  the  restriction  of  fl^-)  to  X  —  .  Finally,  payoff  functions  u-  are  defined 
in  the  obvious  way. 

Now  suppose  that  we  are  given  a  finite-action  approximation  to  a  single  game. 
Since  this  approximation  is  simply  a  family  of  games  parameterised  by  n  e  IN  U  {qd}  and 
satisfying  (Al)-(A3),  its  equilibrium  correspondence  is  upper  semicontinuous  in  n.   Hence, 
since  any  subgame-perfect  equilibrium  path  of  any  game  n  6  IN  is,  in  particular,  an 
equilibrium  path  of  that  game,  it  follows  that  any  limit  point  of  subgame-perfect 
equilibrium  paths  of  the  finite-action  games  is  an  equilibrium  path  of  the  original  game. 
And  our  solution  concept  (namely  subgame-perfect  equilibrium  in  extended  strategies) 
includes  everything  that  can  be  obtained  by  taking  finite-action  approximations  to  a  game 
(or,  more  precisely,  everything  that  can  be  obtained  by  taking  continuous  finite-action 
approximations  to  a  game  satisfying  (Al)-(A3)). 

5.4    Approximation  by  e-equilibria 

One  can  also  show  that  any  limit  point  of  subgame-perfect  e-equilibrium  paths  of 
a  given  single  game  is  an  equilibrium  path  of  the  game,  by  combining  the  techniques  of  this 


39 


paper  with  those  of  Borgers  (1989).   Indeed,  suppose  that  we  are  given  a  family  of  games 
satisfying  (A1)-(A3).   For  each  t  >  1,  xt_1  e  Xt_1   and   e  >  0  ,  let   Et(e,xt_1)  be  the  set 
of  equilibrium  paths  of  subgame  x        of  a  family  of  games  identical  to  the  original  game 
except  that:  (i)  each  player  i  G  Jji  is  replaced  by  a  sequence  {t-|t  >  1}  of  agents;  (ii) 
each  agent  chooses  actions  that  are  at  or  within   e  of  being  optimal  in  each  subgame. 
Then  it  is  easy  to  adapt  the  methods  of  Section  4  to  show  that  E.    is  upper 
semicontinuous.   But  this  implies  the  required  result.   For  any  subgame-perfect 
e— equilibrium  path  of  the  original  game  is  also  an  e-equilibrium  path;  any  e— equilibrium 
path  is  an  e— equilibrium  path  of  the  game  in  which  each  active  player  is  replaced  by  a 
sequence  of  agents;  and  the  0— equilibrium  paths  of  the  game  in  which  each  active  player  is 
replaced  by  a  sequence  of  agents  coincide  with  the  equilibrium  paths  of  the  original  game. 


40 


7    Conclusion 


The  results  of  this  paper  concerning  the  existence  of  subgame-perfect  equilibrium  in 
general  dynamic  games  create  a  small  dilemma.  If  one  adopts  the  point  of  view  that  the 
observation  of  public  signals  violates  the  spirit  of  non-cooperative  game  theory,  then  the 
interpretation  of  the  counterexample  of  Section  2  is  presumably  that  games  with  a 
continuum  of  actions  are  somehow  intrinsically  pathological.   After  all,  that  example  seems 
wholly  non— pathological  in  all  other  respects.  If,  on  the  other  hand,  one  thinks  that  games 
with  a  continuum  of  actions  are  an  indispensable  part  of  non-cooperative  game  theory, 
then  presumably  one  must  accept  that  games  must  be  extended  to  allow  the  observation  of 
public  signals. 

Faced  with  this  dilemma  it  is  tempting  to  dismiss  games  with  a  continuum  of 
actions  as  pathological.   There  are,  however,  at  least  two  reasons  why  this  reaction  may  be 
too  simplistic.   First,  it  can  be  argued  that  a  solution  concept  should  be  robust  to  small 
errors  in  the  data  of  the  games  to  which  it  applies,  in  the  sense  that  one  should  not  rule  out 
equilibria  that  occur  in  games  arbitrarily  close  to  the  game  under  consideration.15  (In  other 
words,  the  equilibrium  correspondence  should  be  upper  semicontinous.)   Provided  that  such 
errors  are  confined  to  payoffs,  the  solution  concept  of  subgame-perfect  equilibrium  satisfies 
this  requirement  when  applied  to  finite— action  games.   If,  however,  such  errors  may  be 
present  in  the  extensive  form  itself,  then  public  signals  must  be  introduced  if  upper 
semicontinuity  is  to  be  satisfied.   Secondly,  it  is  not  entirely  clear  that  public  signals  really 
do  violate  the  spirit  of  non-cooperative  game  theory.   After  all,  players  must  in  any  case 
coordinate  on  a  specific  Nash  equilibrium  to  play;  and  public  signals  merely  increase  the 
extent  to  which  they  can  coordinate,  by  allowing  them  to  coordinate  on  a  random  Nash 
equilibrium. 


15  See  Fudenberg,  Kreps  and  Levine  (1988)  for  example. 


41 


References 


Borgers,  T  (1989):  "Perfect  Equilibrium  Histories  of  Finite—  and  Infinite— Horizon 
Games",  Journal  of  Economic  Theory  47,  218-227. 

Chakrabarti,  S  K  (1988):  "On  the  Existence  of  Equilibria  in  a  Class  of  Discrete-Time 

Dynamic  Games  with  Imperfect  Information",  mimeo,  Department  of  Economics, 
Indiana  University. 

Duffie,  D,  J  Geanakoplos,  A  Mas— Colell  &;  A  McLennan  (1989):  "Stationary  Markov 
Equilibria",  Research  Paper  1079,  Graduate  School  of  Business,  Stanford 
University. 

Forges,  F  (1986):  "An  Approach  to  Communication  Equilibria",  Econometrica  54,  1375- 
1385. 

Fudenberg,  D,  D  Kreps  &  D  Levine  (1988):  "On  the  Robustness  of  Equilibrium 
Refinements",  Journal  of  Economic  Theory  44,  354—380. 

Fudenberg,  D,  &  D  Levine  (1983):  "Subgame-Perfect  Equilibria  of  Finite—  and  Infinite- 
Horizon  Games",  Journal  of  Economic  Theory  31,  251—268. 

Fudenberg,  D,  and  J  Tirole  (1985):  "Pre— Emption  and  Rent  Equalization  in  the  Adoption 
of  a  New  Technology",  Review  of  Economic  Studies  52,  383-401. 

Gihman,  1 1,  and  A  V  Skorohod  (1979):  Controlled  Stochastic  Processes,  Springer— Verlag. 

Harris,  C  (1985a):  "Existence  and  Characterisation  of  Perfect  Equilibrium  in  Games  of 
Perfect  Information",  Econometrica  53,  613-628. 

Harris,  C  (1985b):  "A  Characterisation  of  the  Perfect  Equilibria  of  Infinite-Horizon 
Games",  Journal  of  Economic  Theory  37,  99—125. 

Harris,  C  (1990):  "The  Existence  of  Subgame-Perfect  Equilibrium  with  and  without 

Markov  Strategies:  a  Case  for  Extensive— Form  Correlation",  Discussion  Paper  53, 
Nuffield  College,  Oxford. 

Hellwig,  M,  &;  W  Leininger  (1986):"A  Note  on  the  Relation  between  Discrete  and 

Continuous  Games  of  Perfect  Information",  Discussion  Paper  A— 92,  University  of 
Bonn. 

Hellwig,  M,  &  W  Leininger  (1987):  "On  the  Existence  of  Subgame-Perfect  Equilibrium  in 
Infinite— Action  Games  of  Perfect  Information",  Journal  of  Economic  Theory  43, 
55-75. 

Mertens,  J— F,  &  T  Parthasarathy  (1987):  "Equilibria  for  Discounted  Stochastic  Games", 
Research  Paper  8750,  CORE,  Universite  Catholique  de  Louvain. 

Parthasarathy,  K  R  (1967):  Probability  Measures  on  Metric  Spaces,  Academic  Press. 

Simon,  L  K,  &  W  R  Zame  (1990):  "Discontinuous  Games  and  Endogenous  Sharing 
Rules",  Econometrica  58,  861-872. 


R 


QQ§ 


U 


i 


Date  Due 


MIT    LIBRARIES    DUPL 


3  IDflD  00b72155  t